Guest User

Untitled

a guest
Dec 13th, 2018
98
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 44.33 KB | None | 0 0
  1. name: "Densnet-CIFAR10"
  2.  
  3. layer {
  4. name: "data"
  5. type: "Data"
  6. top: "data"
  7. top: "label"
  8. include {
  9. phase: TRAIN
  10. }
  11. transform_param {
  12. scale: 0.0078125
  13. mirror: true
  14. crop_size: 32
  15. mean_value: 128
  16. }
  17. data_param {
  18. source: "/home/zhibin/caffe/examples/cifar10/cifar10_train_lmdb"
  19. batch_size: 64
  20. backend: LMDB
  21. }
  22. }
  23.  
  24. layer {
  25. name: "data"
  26. type: "Data"
  27. top: "data"
  28. top: "label"
  29. include {
  30. phase: TEST
  31. }
  32. transform_param {
  33. scale: 0.0078125
  34. mirror: true
  35. crop_size: 32
  36. mean_value: 128
  37. }
  38. data_param {
  39. source: "/home/zhibin/caffe/examples/cifar10/cifar10_test_lmdb"
  40. batch_size: 64
  41. backend: LMDB
  42. }
  43. }
  44.  
  45. layer {
  46. name: "data"
  47. type: "Data"
  48. top: "data"
  49. top: "label"
  50. transform_param {
  51. scale: 0.0078125
  52. mirror: true
  53. crop_size: 32
  54. mean_value: 128
  55. }
  56. data_param {
  57. source: "/home/zhibin/caffe/examples/cifar10/cifar10_train_lmdb"
  58. batch_size: 64
  59. backend: LMDB
  60. }
  61. }
  62. layer {
  63. name: "Convolution1"
  64. type: "Convolution"
  65. bottom: "data"
  66. top: "Convolution1"
  67. convolution_param {
  68. num_output: 16
  69. bias_term: false
  70. pad: 1
  71. kernel_size: 3
  72. stride: 1
  73. weight_filler {
  74. type: "msra"
  75. }
  76. bias_filler {
  77. type: "constant"
  78. }
  79. }
  80. }
  81. layer {
  82. name: "BatchNorm1"
  83. type: "BatchNorm"
  84. bottom: "Convolution1"
  85. top: "BatchNorm1"
  86. param {
  87. lr_mult: 0.0
  88. decay_mult: 0.0
  89. }
  90. param {
  91. lr_mult: 0.0
  92. decay_mult: 0.0
  93. }
  94. param {
  95. lr_mult: 0.0
  96. decay_mult: 0.0
  97. }
  98. }
  99. layer {
  100. name: "Scale1"
  101. type: "Scale"
  102. bottom: "BatchNorm1"
  103. top: "BatchNorm1"
  104. scale_param {
  105. filler {
  106. value: 1.0
  107. }
  108. bias_term: true
  109. bias_filler {
  110. value: 0.0
  111. }
  112. }
  113. }
  114. layer {
  115. name: "ReLU1"
  116. type: "ReLU"
  117. bottom: "BatchNorm1"
  118. top: "BatchNorm1"
  119. }
  120. layer {
  121. name: "Convolution2"
  122. type: "Convolution"
  123. bottom: "BatchNorm1"
  124. top: "Convolution2"
  125. convolution_param {
  126. num_output: 12
  127. bias_term: false
  128. pad: 1
  129. kernel_size: 3
  130. stride: 1
  131. weight_filler {
  132. type: "msra"
  133. }
  134. bias_filler {
  135. type: "constant"
  136. }
  137. }
  138. }
  139. layer {
  140. name: "Dropout1"
  141. type: "Dropout"
  142. bottom: "Convolution2"
  143. top: "Dropout1"
  144. dropout_param {
  145. dropout_ratio: 0.20000000298
  146. }
  147. }
  148. layer {
  149. name: "Concat1"
  150. type: "Concat"
  151. bottom: "Convolution1"
  152. bottom: "Dropout1"
  153. top: "Concat1"
  154. concat_param {
  155. axis: 1
  156. }
  157. }
  158. layer {
  159. name: "BatchNorm2"
  160. type: "BatchNorm"
  161. bottom: "Concat1"
  162. top: "BatchNorm2"
  163. param {
  164. lr_mult: 0.0
  165. decay_mult: 0.0
  166. }
  167. param {
  168. lr_mult: 0.0
  169. decay_mult: 0.0
  170. }
  171. param {
  172. lr_mult: 0.0
  173. decay_mult: 0.0
  174. }
  175. }
  176. layer {
  177. name: "Scale2"
  178. type: "Scale"
  179. bottom: "BatchNorm2"
  180. top: "BatchNorm2"
  181. scale_param {
  182. filler {
  183. value: 1.0
  184. }
  185. bias_term: true
  186. bias_filler {
  187. value: 0.0
  188. }
  189. }
  190. }
  191. layer {
  192. name: "ReLU2"
  193. type: "ReLU"
  194. bottom: "BatchNorm2"
  195. top: "BatchNorm2"
  196. }
  197. layer {
  198. name: "Convolution3"
  199. type: "Convolution"
  200. bottom: "BatchNorm2"
  201. top: "Convolution3"
  202. convolution_param {
  203. num_output: 12
  204. bias_term: false
  205. pad: 1
  206. kernel_size: 3
  207. stride: 1
  208. weight_filler {
  209. type: "msra"
  210. }
  211. bias_filler {
  212. type: "constant"
  213. }
  214. }
  215. }
  216. layer {
  217. name: "Dropout2"
  218. type: "Dropout"
  219. bottom: "Convolution3"
  220. top: "Dropout2"
  221. dropout_param {
  222. dropout_ratio: 0.20000000298
  223. }
  224. }
  225. layer {
  226. name: "Concat2"
  227. type: "Concat"
  228. bottom: "Concat1"
  229. bottom: "Dropout2"
  230. top: "Concat2"
  231. concat_param {
  232. axis: 1
  233. }
  234. }
  235. layer {
  236. name: "BatchNorm3"
  237. type: "BatchNorm"
  238. bottom: "Concat2"
  239. top: "BatchNorm3"
  240. param {
  241. lr_mult: 0.0
  242. decay_mult: 0.0
  243. }
  244. param {
  245. lr_mult: 0.0
  246. decay_mult: 0.0
  247. }
  248. param {
  249. lr_mult: 0.0
  250. decay_mult: 0.0
  251. }
  252. }
  253. layer {
  254. name: "Scale3"
  255. type: "Scale"
  256. bottom: "BatchNorm3"
  257. top: "BatchNorm3"
  258. scale_param {
  259. filler {
  260. value: 1.0
  261. }
  262. bias_term: true
  263. bias_filler {
  264. value: 0.0
  265. }
  266. }
  267. }
  268. layer {
  269. name: "ReLU3"
  270. type: "ReLU"
  271. bottom: "BatchNorm3"
  272. top: "BatchNorm3"
  273. }
  274. layer {
  275. name: "Convolution4"
  276. type: "Convolution"
  277. bottom: "BatchNorm3"
  278. top: "Convolution4"
  279. convolution_param {
  280. num_output: 12
  281. bias_term: false
  282. pad: 1
  283. kernel_size: 3
  284. stride: 1
  285. weight_filler {
  286. type: "msra"
  287. }
  288. bias_filler {
  289. type: "constant"
  290. }
  291. }
  292. }
  293. layer {
  294. name: "Dropout3"
  295. type: "Dropout"
  296. bottom: "Convolution4"
  297. top: "Dropout3"
  298. dropout_param {
  299. dropout_ratio: 0.20000000298
  300. }
  301. }
  302. layer {
  303. name: "Concat3"
  304. type: "Concat"
  305. bottom: "Concat2"
  306. bottom: "Dropout3"
  307. top: "Concat3"
  308. concat_param {
  309. axis: 1
  310. }
  311. }
  312. layer {
  313. name: "BatchNorm4"
  314. type: "BatchNorm"
  315. bottom: "Concat3"
  316. top: "BatchNorm4"
  317. param {
  318. lr_mult: 0.0
  319. decay_mult: 0.0
  320. }
  321. param {
  322. lr_mult: 0.0
  323. decay_mult: 0.0
  324. }
  325. param {
  326. lr_mult: 0.0
  327. decay_mult: 0.0
  328. }
  329. }
  330. layer {
  331. name: "Scale4"
  332. type: "Scale"
  333. bottom: "BatchNorm4"
  334. top: "BatchNorm4"
  335. scale_param {
  336. filler {
  337. value: 1.0
  338. }
  339. bias_term: true
  340. bias_filler {
  341. value: 0.0
  342. }
  343. }
  344. }
  345. layer {
  346. name: "ReLU4"
  347. type: "ReLU"
  348. bottom: "BatchNorm4"
  349. top: "BatchNorm4"
  350. }
  351. layer {
  352. name: "Convolution5"
  353. type: "Convolution"
  354. bottom: "BatchNorm4"
  355. top: "Convolution5"
  356. convolution_param {
  357. num_output: 12
  358. bias_term: false
  359. pad: 1
  360. kernel_size: 3
  361. stride: 1
  362. weight_filler {
  363. type: "msra"
  364. }
  365. bias_filler {
  366. type: "constant"
  367. }
  368. }
  369. }
  370. layer {
  371. name: "Dropout4"
  372. type: "Dropout"
  373. bottom: "Convolution5"
  374. top: "Dropout4"
  375. dropout_param {
  376. dropout_ratio: 0.20000000298
  377. }
  378. }
  379. layer {
  380. name: "Concat4"
  381. type: "Concat"
  382. bottom: "Concat3"
  383. bottom: "Dropout4"
  384. top: "Concat4"
  385. concat_param {
  386. axis: 1
  387. }
  388. }
  389. layer {
  390. name: "BatchNorm5"
  391. type: "BatchNorm"
  392. bottom: "Concat4"
  393. top: "BatchNorm5"
  394. param {
  395. lr_mult: 0.0
  396. decay_mult: 0.0
  397. }
  398. param {
  399. lr_mult: 0.0
  400. decay_mult: 0.0
  401. }
  402. param {
  403. lr_mult: 0.0
  404. decay_mult: 0.0
  405. }
  406. }
  407. layer {
  408. name: "Scale5"
  409. type: "Scale"
  410. bottom: "BatchNorm5"
  411. top: "BatchNorm5"
  412. scale_param {
  413. filler {
  414. value: 1.0
  415. }
  416. bias_term: true
  417. bias_filler {
  418. value: 0.0
  419. }
  420. }
  421. }
  422. layer {
  423. name: "ReLU5"
  424. type: "ReLU"
  425. bottom: "BatchNorm5"
  426. top: "BatchNorm5"
  427. }
  428. layer {
  429. name: "Convolution6"
  430. type: "Convolution"
  431. bottom: "BatchNorm5"
  432. top: "Convolution6"
  433. convolution_param {
  434. num_output: 12
  435. bias_term: false
  436. pad: 1
  437. kernel_size: 3
  438. stride: 1
  439. weight_filler {
  440. type: "msra"
  441. }
  442. bias_filler {
  443. type: "constant"
  444. }
  445. }
  446. }
  447. layer {
  448. name: "Dropout5"
  449. type: "Dropout"
  450. bottom: "Convolution6"
  451. top: "Dropout5"
  452. dropout_param {
  453. dropout_ratio: 0.20000000298
  454. }
  455. }
  456. layer {
  457. name: "Concat5"
  458. type: "Concat"
  459. bottom: "Concat4"
  460. bottom: "Dropout5"
  461. top: "Concat5"
  462. concat_param {
  463. axis: 1
  464. }
  465. }
  466. layer {
  467. name: "BatchNorm6"
  468. type: "BatchNorm"
  469. bottom: "Concat5"
  470. top: "BatchNorm6"
  471. param {
  472. lr_mult: 0.0
  473. decay_mult: 0.0
  474. }
  475. param {
  476. lr_mult: 0.0
  477. decay_mult: 0.0
  478. }
  479. param {
  480. lr_mult: 0.0
  481. decay_mult: 0.0
  482. }
  483. }
  484. layer {
  485. name: "Scale6"
  486. type: "Scale"
  487. bottom: "BatchNorm6"
  488. top: "BatchNorm6"
  489. scale_param {
  490. filler {
  491. value: 1.0
  492. }
  493. bias_term: true
  494. bias_filler {
  495. value: 0.0
  496. }
  497. }
  498. }
  499. layer {
  500. name: "ReLU6"
  501. type: "ReLU"
  502. bottom: "BatchNorm6"
  503. top: "BatchNorm6"
  504. }
  505. layer {
  506. name: "Convolution7"
  507. type: "Convolution"
  508. bottom: "BatchNorm6"
  509. top: "Convolution7"
  510. convolution_param {
  511. num_output: 12
  512. bias_term: false
  513. pad: 1
  514. kernel_size: 3
  515. stride: 1
  516. weight_filler {
  517. type: "msra"
  518. }
  519. bias_filler {
  520. type: "constant"
  521. }
  522. }
  523. }
  524. layer {
  525. name: "Dropout6"
  526. type: "Dropout"
  527. bottom: "Convolution7"
  528. top: "Dropout6"
  529. dropout_param {
  530. dropout_ratio: 0.20000000298
  531. }
  532. }
  533. layer {
  534. name: "Concat6"
  535. type: "Concat"
  536. bottom: "Concat5"
  537. bottom: "Dropout6"
  538. top: "Concat6"
  539. concat_param {
  540. axis: 1
  541. }
  542. }
  543. layer {
  544. name: "BatchNorm7"
  545. type: "BatchNorm"
  546. bottom: "Concat6"
  547. top: "BatchNorm7"
  548. param {
  549. lr_mult: 0.0
  550. decay_mult: 0.0
  551. }
  552. param {
  553. lr_mult: 0.0
  554. decay_mult: 0.0
  555. }
  556. param {
  557. lr_mult: 0.0
  558. decay_mult: 0.0
  559. }
  560. }
  561. layer {
  562. name: "Scale7"
  563. type: "Scale"
  564. bottom: "BatchNorm7"
  565. top: "BatchNorm7"
  566. scale_param {
  567. filler {
  568. value: 1.0
  569. }
  570. bias_term: true
  571. bias_filler {
  572. value: 0.0
  573. }
  574. }
  575. }
  576. layer {
  577. name: "ReLU7"
  578. type: "ReLU"
  579. bottom: "BatchNorm7"
  580. top: "BatchNorm7"
  581. }
  582. layer {
  583. name: "Convolution8"
  584. type: "Convolution"
  585. bottom: "BatchNorm7"
  586. top: "Convolution8"
  587. convolution_param {
  588. num_output: 12
  589. bias_term: false
  590. pad: 1
  591. kernel_size: 3
  592. stride: 1
  593. weight_filler {
  594. type: "msra"
  595. }
  596. bias_filler {
  597. type: "constant"
  598. }
  599. }
  600. }
  601. layer {
  602. name: "Dropout7"
  603. type: "Dropout"
  604. bottom: "Convolution8"
  605. top: "Dropout7"
  606. dropout_param {
  607. dropout_ratio: 0.20000000298
  608. }
  609. }
  610. layer {
  611. name: "Concat7"
  612. type: "Concat"
  613. bottom: "Concat6"
  614. bottom: "Dropout7"
  615. top: "Concat7"
  616. concat_param {
  617. axis: 1
  618. }
  619. }
  620. layer {
  621. name: "BatchNorm8"
  622. type: "BatchNorm"
  623. bottom: "Concat7"
  624. top: "BatchNorm8"
  625. param {
  626. lr_mult: 0.0
  627. decay_mult: 0.0
  628. }
  629. param {
  630. lr_mult: 0.0
  631. decay_mult: 0.0
  632. }
  633. param {
  634. lr_mult: 0.0
  635. decay_mult: 0.0
  636. }
  637. }
  638. layer {
  639. name: "Scale8"
  640. type: "Scale"
  641. bottom: "BatchNorm8"
  642. top: "BatchNorm8"
  643. scale_param {
  644. filler {
  645. value: 1.0
  646. }
  647. bias_term: true
  648. bias_filler {
  649. value: 0.0
  650. }
  651. }
  652. }
  653. layer {
  654. name: "ReLU8"
  655. type: "ReLU"
  656. bottom: "BatchNorm8"
  657. top: "BatchNorm8"
  658. }
  659. layer {
  660. name: "Convolution9"
  661. type: "Convolution"
  662. bottom: "BatchNorm8"
  663. top: "Convolution9"
  664. convolution_param {
  665. num_output: 12
  666. bias_term: false
  667. pad: 1
  668. kernel_size: 3
  669. stride: 1
  670. weight_filler {
  671. type: "msra"
  672. }
  673. bias_filler {
  674. type: "constant"
  675. }
  676. }
  677. }
  678. layer {
  679. name: "Dropout8"
  680. type: "Dropout"
  681. bottom: "Convolution9"
  682. top: "Dropout8"
  683. dropout_param {
  684. dropout_ratio: 0.20000000298
  685. }
  686. }
  687. layer {
  688. name: "Concat8"
  689. type: "Concat"
  690. bottom: "Concat7"
  691. bottom: "Dropout8"
  692. top: "Concat8"
  693. concat_param {
  694. axis: 1
  695. }
  696. }
  697. layer {
  698. name: "BatchNorm9"
  699. type: "BatchNorm"
  700. bottom: "Concat8"
  701. top: "BatchNorm9"
  702. param {
  703. lr_mult: 0.0
  704. decay_mult: 0.0
  705. }
  706. param {
  707. lr_mult: 0.0
  708. decay_mult: 0.0
  709. }
  710. param {
  711. lr_mult: 0.0
  712. decay_mult: 0.0
  713. }
  714. }
  715. layer {
  716. name: "Scale9"
  717. type: "Scale"
  718. bottom: "BatchNorm9"
  719. top: "BatchNorm9"
  720. scale_param {
  721. filler {
  722. value: 1.0
  723. }
  724. bias_term: true
  725. bias_filler {
  726. value: 0.0
  727. }
  728. }
  729. }
  730. layer {
  731. name: "ReLU9"
  732. type: "ReLU"
  733. bottom: "BatchNorm9"
  734. top: "BatchNorm9"
  735. }
  736. layer {
  737. name: "Convolution10"
  738. type: "Convolution"
  739. bottom: "BatchNorm9"
  740. top: "Convolution10"
  741. convolution_param {
  742. num_output: 12
  743. bias_term: false
  744. pad: 1
  745. kernel_size: 3
  746. stride: 1
  747. weight_filler {
  748. type: "msra"
  749. }
  750. bias_filler {
  751. type: "constant"
  752. }
  753. }
  754. }
  755. layer {
  756. name: "Dropout9"
  757. type: "Dropout"
  758. bottom: "Convolution10"
  759. top: "Dropout9"
  760. dropout_param {
  761. dropout_ratio: 0.20000000298
  762. }
  763. }
  764. layer {
  765. name: "Concat9"
  766. type: "Concat"
  767. bottom: "Concat8"
  768. bottom: "Dropout9"
  769. top: "Concat9"
  770. concat_param {
  771. axis: 1
  772. }
  773. }
  774. layer {
  775. name: "BatchNorm10"
  776. type: "BatchNorm"
  777. bottom: "Concat9"
  778. top: "BatchNorm10"
  779. param {
  780. lr_mult: 0.0
  781. decay_mult: 0.0
  782. }
  783. param {
  784. lr_mult: 0.0
  785. decay_mult: 0.0
  786. }
  787. param {
  788. lr_mult: 0.0
  789. decay_mult: 0.0
  790. }
  791. }
  792. layer {
  793. name: "Scale10"
  794. type: "Scale"
  795. bottom: "BatchNorm10"
  796. top: "BatchNorm10"
  797. scale_param {
  798. filler {
  799. value: 1.0
  800. }
  801. bias_term: true
  802. bias_filler {
  803. value: 0.0
  804. }
  805. }
  806. }
  807. layer {
  808. name: "ReLU10"
  809. type: "ReLU"
  810. bottom: "BatchNorm10"
  811. top: "BatchNorm10"
  812. }
  813. layer {
  814. name: "Convolution11"
  815. type: "Convolution"
  816. bottom: "BatchNorm10"
  817. top: "Convolution11"
  818. convolution_param {
  819. num_output: 12
  820. bias_term: false
  821. pad: 1
  822. kernel_size: 3
  823. stride: 1
  824. weight_filler {
  825. type: "msra"
  826. }
  827. bias_filler {
  828. type: "constant"
  829. }
  830. }
  831. }
  832. layer {
  833. name: "Dropout10"
  834. type: "Dropout"
  835. bottom: "Convolution11"
  836. top: "Dropout10"
  837. dropout_param {
  838. dropout_ratio: 0.20000000298
  839. }
  840. }
  841. layer {
  842. name: "Concat10"
  843. type: "Concat"
  844. bottom: "Concat9"
  845. bottom: "Dropout10"
  846. top: "Concat10"
  847. concat_param {
  848. axis: 1
  849. }
  850. }
  851. layer {
  852. name: "BatchNorm11"
  853. type: "BatchNorm"
  854. bottom: "Concat10"
  855. top: "BatchNorm11"
  856. param {
  857. lr_mult: 0.0
  858. decay_mult: 0.0
  859. }
  860. param {
  861. lr_mult: 0.0
  862. decay_mult: 0.0
  863. }
  864. param {
  865. lr_mult: 0.0
  866. decay_mult: 0.0
  867. }
  868. }
  869. layer {
  870. name: "Scale11"
  871. type: "Scale"
  872. bottom: "BatchNorm11"
  873. top: "BatchNorm11"
  874. scale_param {
  875. filler {
  876. value: 1.0
  877. }
  878. bias_term: true
  879. bias_filler {
  880. value: 0.0
  881. }
  882. }
  883. }
  884. layer {
  885. name: "ReLU11"
  886. type: "ReLU"
  887. bottom: "BatchNorm11"
  888. top: "BatchNorm11"
  889. }
  890. layer {
  891. name: "Convolution12"
  892. type: "Convolution"
  893. bottom: "BatchNorm11"
  894. top: "Convolution12"
  895. convolution_param {
  896. num_output: 12
  897. bias_term: false
  898. pad: 1
  899. kernel_size: 3
  900. stride: 1
  901. weight_filler {
  902. type: "msra"
  903. }
  904. bias_filler {
  905. type: "constant"
  906. }
  907. }
  908. }
  909. layer {
  910. name: "Dropout11"
  911. type: "Dropout"
  912. bottom: "Convolution12"
  913. top: "Dropout11"
  914. dropout_param {
  915. dropout_ratio: 0.20000000298
  916. }
  917. }
  918. layer {
  919. name: "Concat11"
  920. type: "Concat"
  921. bottom: "Concat10"
  922. bottom: "Dropout11"
  923. top: "Concat11"
  924. concat_param {
  925. axis: 1
  926. }
  927. }
  928. layer {
  929. name: "BatchNorm12"
  930. type: "BatchNorm"
  931. bottom: "Concat11"
  932. top: "BatchNorm12"
  933. param {
  934. lr_mult: 0.0
  935. decay_mult: 0.0
  936. }
  937. param {
  938. lr_mult: 0.0
  939. decay_mult: 0.0
  940. }
  941. param {
  942. lr_mult: 0.0
  943. decay_mult: 0.0
  944. }
  945. }
  946. layer {
  947. name: "Scale12"
  948. type: "Scale"
  949. bottom: "BatchNorm12"
  950. top: "BatchNorm12"
  951. scale_param {
  952. filler {
  953. value: 1.0
  954. }
  955. bias_term: true
  956. bias_filler {
  957. value: 0.0
  958. }
  959. }
  960. }
  961. layer {
  962. name: "ReLU12"
  963. type: "ReLU"
  964. bottom: "BatchNorm12"
  965. top: "BatchNorm12"
  966. }
  967. layer {
  968. name: "Convolution13"
  969. type: "Convolution"
  970. bottom: "BatchNorm12"
  971. top: "Convolution13"
  972. convolution_param {
  973. num_output: 12
  974. bias_term: false
  975. pad: 1
  976. kernel_size: 3
  977. stride: 1
  978. weight_filler {
  979. type: "msra"
  980. }
  981. bias_filler {
  982. type: "constant"
  983. }
  984. }
  985. }
  986. layer {
  987. name: "Dropout12"
  988. type: "Dropout"
  989. bottom: "Convolution13"
  990. top: "Dropout12"
  991. dropout_param {
  992. dropout_ratio: 0.20000000298
  993. }
  994. }
  995. layer {
  996. name: "Concat12"
  997. type: "Concat"
  998. bottom: "Concat11"
  999. bottom: "Dropout12"
  1000. top: "Concat12"
  1001. concat_param {
  1002. axis: 1
  1003. }
  1004. }
  1005. layer {
  1006. name: "BatchNorm13"
  1007. type: "BatchNorm"
  1008. bottom: "Concat12"
  1009. top: "BatchNorm13"
  1010. param {
  1011. lr_mult: 0.0
  1012. decay_mult: 0.0
  1013. }
  1014. param {
  1015. lr_mult: 0.0
  1016. decay_mult: 0.0
  1017. }
  1018. param {
  1019. lr_mult: 0.0
  1020. decay_mult: 0.0
  1021. }
  1022. }
  1023. layer {
  1024. name: "Scale13"
  1025. type: "Scale"
  1026. bottom: "BatchNorm13"
  1027. top: "BatchNorm13"
  1028. scale_param {
  1029. filler {
  1030. value: 1.0
  1031. }
  1032. bias_term: true
  1033. bias_filler {
  1034. value: 0.0
  1035. }
  1036. }
  1037. }
  1038. layer {
  1039. name: "ReLU13"
  1040. type: "ReLU"
  1041. bottom: "BatchNorm13"
  1042. top: "BatchNorm13"
  1043. }
  1044. layer {
  1045. name: "Convolution14"
  1046. type: "Convolution"
  1047. bottom: "BatchNorm13"
  1048. top: "Convolution14"
  1049. convolution_param {
  1050. num_output: 160
  1051. bias_term: false
  1052. pad: 0
  1053. kernel_size: 1
  1054. stride: 1
  1055. weight_filler {
  1056. type: "msra"
  1057. }
  1058. bias_filler {
  1059. type: "constant"
  1060. }
  1061. }
  1062. }
  1063. layer {
  1064. name: "Dropout13"
  1065. type: "Dropout"
  1066. bottom: "Convolution14"
  1067. top: "Dropout13"
  1068. dropout_param {
  1069. dropout_ratio: 0.20000000298
  1070. }
  1071. }
  1072. layer {
  1073. name: "Pooling1"
  1074. type: "Pooling"
  1075. bottom: "Dropout13"
  1076. top: "Pooling1"
  1077. pooling_param {
  1078. pool: AVE
  1079. kernel_size: 2
  1080. stride: 2
  1081. }
  1082. }
  1083. layer {
  1084. name: "BatchNorm14"
  1085. type: "BatchNorm"
  1086. bottom: "Pooling1"
  1087. top: "BatchNorm14"
  1088. param {
  1089. lr_mult: 0.0
  1090. decay_mult: 0.0
  1091. }
  1092. param {
  1093. lr_mult: 0.0
  1094. decay_mult: 0.0
  1095. }
  1096. param {
  1097. lr_mult: 0.0
  1098. decay_mult: 0.0
  1099. }
  1100. }
  1101. layer {
  1102. name: "Scale14"
  1103. type: "Scale"
  1104. bottom: "BatchNorm14"
  1105. top: "BatchNorm14"
  1106. scale_param {
  1107. filler {
  1108. value: 1.0
  1109. }
  1110. bias_term: true
  1111. bias_filler {
  1112. value: 0.0
  1113. }
  1114. }
  1115. }
  1116. layer {
  1117. name: "ReLU14"
  1118. type: "ReLU"
  1119. bottom: "BatchNorm14"
  1120. top: "BatchNorm14"
  1121. }
  1122. layer {
  1123. name: "Convolution15"
  1124. type: "Convolution"
  1125. bottom: "BatchNorm14"
  1126. top: "Convolution15"
  1127. convolution_param {
  1128. num_output: 12
  1129. bias_term: false
  1130. pad: 1
  1131. kernel_size: 3
  1132. stride: 1
  1133. weight_filler {
  1134. type: "msra"
  1135. }
  1136. bias_filler {
  1137. type: "constant"
  1138. }
  1139. }
  1140. }
  1141. layer {
  1142. name: "Dropout14"
  1143. type: "Dropout"
  1144. bottom: "Convolution15"
  1145. top: "Dropout14"
  1146. dropout_param {
  1147. dropout_ratio: 0.20000000298
  1148. }
  1149. }
  1150. layer {
  1151. name: "Concat13"
  1152. type: "Concat"
  1153. bottom: "Pooling1"
  1154. bottom: "Dropout14"
  1155. top: "Concat13"
  1156. concat_param {
  1157. axis: 1
  1158. }
  1159. }
  1160. layer {
  1161. name: "BatchNorm15"
  1162. type: "BatchNorm"
  1163. bottom: "Concat13"
  1164. top: "BatchNorm15"
  1165. param {
  1166. lr_mult: 0.0
  1167. decay_mult: 0.0
  1168. }
  1169. param {
  1170. lr_mult: 0.0
  1171. decay_mult: 0.0
  1172. }
  1173. param {
  1174. lr_mult: 0.0
  1175. decay_mult: 0.0
  1176. }
  1177. }
  1178. layer {
  1179. name: "Scale15"
  1180. type: "Scale"
  1181. bottom: "BatchNorm15"
  1182. top: "BatchNorm15"
  1183. scale_param {
  1184. filler {
  1185. value: 1.0
  1186. }
  1187. bias_term: true
  1188. bias_filler {
  1189. value: 0.0
  1190. }
  1191. }
  1192. }
  1193. layer {
  1194. name: "ReLU15"
  1195. type: "ReLU"
  1196. bottom: "BatchNorm15"
  1197. top: "BatchNorm15"
  1198. }
  1199. layer {
  1200. name: "Convolution16"
  1201. type: "Convolution"
  1202. bottom: "BatchNorm15"
  1203. top: "Convolution16"
  1204. convolution_param {
  1205. num_output: 12
  1206. bias_term: false
  1207. pad: 1
  1208. kernel_size: 3
  1209. stride: 1
  1210. weight_filler {
  1211. type: "msra"
  1212. }
  1213. bias_filler {
  1214. type: "constant"
  1215. }
  1216. }
  1217. }
  1218. layer {
  1219. name: "Dropout15"
  1220. type: "Dropout"
  1221. bottom: "Convolution16"
  1222. top: "Dropout15"
  1223. dropout_param {
  1224. dropout_ratio: 0.20000000298
  1225. }
  1226. }
  1227. layer {
  1228. name: "Concat14"
  1229. type: "Concat"
  1230. bottom: "Concat13"
  1231. bottom: "Dropout15"
  1232. top: "Concat14"
  1233. concat_param {
  1234. axis: 1
  1235. }
  1236. }
  1237. layer {
  1238. name: "BatchNorm16"
  1239. type: "BatchNorm"
  1240. bottom: "Concat14"
  1241. top: "BatchNorm16"
  1242. param {
  1243. lr_mult: 0.0
  1244. decay_mult: 0.0
  1245. }
  1246. param {
  1247. lr_mult: 0.0
  1248. decay_mult: 0.0
  1249. }
  1250. param {
  1251. lr_mult: 0.0
  1252. decay_mult: 0.0
  1253. }
  1254. }
  1255. layer {
  1256. name: "Scale16"
  1257. type: "Scale"
  1258. bottom: "BatchNorm16"
  1259. top: "BatchNorm16"
  1260. scale_param {
  1261. filler {
  1262. value: 1.0
  1263. }
  1264. bias_term: true
  1265. bias_filler {
  1266. value: 0.0
  1267. }
  1268. }
  1269. }
  1270. layer {
  1271. name: "ReLU16"
  1272. type: "ReLU"
  1273. bottom: "BatchNorm16"
  1274. top: "BatchNorm16"
  1275. }
  1276. layer {
  1277. name: "Convolution17"
  1278. type: "Convolution"
  1279. bottom: "BatchNorm16"
  1280. top: "Convolution17"
  1281. convolution_param {
  1282. num_output: 12
  1283. bias_term: false
  1284. pad: 1
  1285. kernel_size: 3
  1286. stride: 1
  1287. weight_filler {
  1288. type: "msra"
  1289. }
  1290. bias_filler {
  1291. type: "constant"
  1292. }
  1293. }
  1294. }
  1295. layer {
  1296. name: "Dropout16"
  1297. type: "Dropout"
  1298. bottom: "Convolution17"
  1299. top: "Dropout16"
  1300. dropout_param {
  1301. dropout_ratio: 0.20000000298
  1302. }
  1303. }
  1304. layer {
  1305. name: "Concat15"
  1306. type: "Concat"
  1307. bottom: "Concat14"
  1308. bottom: "Dropout16"
  1309. top: "Concat15"
  1310. concat_param {
  1311. axis: 1
  1312. }
  1313. }
  1314. layer {
  1315. name: "BatchNorm17"
  1316. type: "BatchNorm"
  1317. bottom: "Concat15"
  1318. top: "BatchNorm17"
  1319. param {
  1320. lr_mult: 0.0
  1321. decay_mult: 0.0
  1322. }
  1323. param {
  1324. lr_mult: 0.0
  1325. decay_mult: 0.0
  1326. }
  1327. param {
  1328. lr_mult: 0.0
  1329. decay_mult: 0.0
  1330. }
  1331. }
  1332. layer {
  1333. name: "Scale17"
  1334. type: "Scale"
  1335. bottom: "BatchNorm17"
  1336. top: "BatchNorm17"
  1337. scale_param {
  1338. filler {
  1339. value: 1.0
  1340. }
  1341. bias_term: true
  1342. bias_filler {
  1343. value: 0.0
  1344. }
  1345. }
  1346. }
  1347. layer {
  1348. name: "ReLU17"
  1349. type: "ReLU"
  1350. bottom: "BatchNorm17"
  1351. top: "BatchNorm17"
  1352. }
  1353. layer {
  1354. name: "Convolution18"
  1355. type: "Convolution"
  1356. bottom: "BatchNorm17"
  1357. top: "Convolution18"
  1358. convolution_param {
  1359. num_output: 12
  1360. bias_term: false
  1361. pad: 1
  1362. kernel_size: 3
  1363. stride: 1
  1364. weight_filler {
  1365. type: "msra"
  1366. }
  1367. bias_filler {
  1368. type: "constant"
  1369. }
  1370. }
  1371. }
  1372. layer {
  1373. name: "Dropout17"
  1374. type: "Dropout"
  1375. bottom: "Convolution18"
  1376. top: "Dropout17"
  1377. dropout_param {
  1378. dropout_ratio: 0.20000000298
  1379. }
  1380. }
  1381. layer {
  1382. name: "Concat16"
  1383. type: "Concat"
  1384. bottom: "Concat15"
  1385. bottom: "Dropout17"
  1386. top: "Concat16"
  1387. concat_param {
  1388. axis: 1
  1389. }
  1390. }
  1391. layer {
  1392. name: "BatchNorm18"
  1393. type: "BatchNorm"
  1394. bottom: "Concat16"
  1395. top: "BatchNorm18"
  1396. param {
  1397. lr_mult: 0.0
  1398. decay_mult: 0.0
  1399. }
  1400. param {
  1401. lr_mult: 0.0
  1402. decay_mult: 0.0
  1403. }
  1404. param {
  1405. lr_mult: 0.0
  1406. decay_mult: 0.0
  1407. }
  1408. }
  1409. layer {
  1410. name: "Scale18"
  1411. type: "Scale"
  1412. bottom: "BatchNorm18"
  1413. top: "BatchNorm18"
  1414. scale_param {
  1415. filler {
  1416. value: 1.0
  1417. }
  1418. bias_term: true
  1419. bias_filler {
  1420. value: 0.0
  1421. }
  1422. }
  1423. }
  1424. layer {
  1425. name: "ReLU18"
  1426. type: "ReLU"
  1427. bottom: "BatchNorm18"
  1428. top: "BatchNorm18"
  1429. }
  1430. layer {
  1431. name: "Convolution19"
  1432. type: "Convolution"
  1433. bottom: "BatchNorm18"
  1434. top: "Convolution19"
  1435. convolution_param {
  1436. num_output: 12
  1437. bias_term: false
  1438. pad: 1
  1439. kernel_size: 3
  1440. stride: 1
  1441. weight_filler {
  1442. type: "msra"
  1443. }
  1444. bias_filler {
  1445. type: "constant"
  1446. }
  1447. }
  1448. }
  1449. layer {
  1450. name: "Dropout18"
  1451. type: "Dropout"
  1452. bottom: "Convolution19"
  1453. top: "Dropout18"
  1454. dropout_param {
  1455. dropout_ratio: 0.20000000298
  1456. }
  1457. }
  1458. layer {
  1459. name: "Concat17"
  1460. type: "Concat"
  1461. bottom: "Concat16"
  1462. bottom: "Dropout18"
  1463. top: "Concat17"
  1464. concat_param {
  1465. axis: 1
  1466. }
  1467. }
  1468. layer {
  1469. name: "BatchNorm19"
  1470. type: "BatchNorm"
  1471. bottom: "Concat17"
  1472. top: "BatchNorm19"
  1473. param {
  1474. lr_mult: 0.0
  1475. decay_mult: 0.0
  1476. }
  1477. param {
  1478. lr_mult: 0.0
  1479. decay_mult: 0.0
  1480. }
  1481. param {
  1482. lr_mult: 0.0
  1483. decay_mult: 0.0
  1484. }
  1485. }
  1486. layer {
  1487. name: "Scale19"
  1488. type: "Scale"
  1489. bottom: "BatchNorm19"
  1490. top: "BatchNorm19"
  1491. scale_param {
  1492. filler {
  1493. value: 1.0
  1494. }
  1495. bias_term: true
  1496. bias_filler {
  1497. value: 0.0
  1498. }
  1499. }
  1500. }
  1501. layer {
  1502. name: "ReLU19"
  1503. type: "ReLU"
  1504. bottom: "BatchNorm19"
  1505. top: "BatchNorm19"
  1506. }
  1507. layer {
  1508. name: "Convolution20"
  1509. type: "Convolution"
  1510. bottom: "BatchNorm19"
  1511. top: "Convolution20"
  1512. convolution_param {
  1513. num_output: 12
  1514. bias_term: false
  1515. pad: 1
  1516. kernel_size: 3
  1517. stride: 1
  1518. weight_filler {
  1519. type: "msra"
  1520. }
  1521. bias_filler {
  1522. type: "constant"
  1523. }
  1524. }
  1525. }
  1526. layer {
  1527. name: "Dropout19"
  1528. type: "Dropout"
  1529. bottom: "Convolution20"
  1530. top: "Dropout19"
  1531. dropout_param {
  1532. dropout_ratio: 0.20000000298
  1533. }
  1534. }
  1535. layer {
  1536. name: "Concat18"
  1537. type: "Concat"
  1538. bottom: "Concat17"
  1539. bottom: "Dropout19"
  1540. top: "Concat18"
  1541. concat_param {
  1542. axis: 1
  1543. }
  1544. }
  1545. layer {
  1546. name: "BatchNorm20"
  1547. type: "BatchNorm"
  1548. bottom: "Concat18"
  1549. top: "BatchNorm20"
  1550. param {
  1551. lr_mult: 0.0
  1552. decay_mult: 0.0
  1553. }
  1554. param {
  1555. lr_mult: 0.0
  1556. decay_mult: 0.0
  1557. }
  1558. param {
  1559. lr_mult: 0.0
  1560. decay_mult: 0.0
  1561. }
  1562. }
  1563. layer {
  1564. name: "Scale20"
  1565. type: "Scale"
  1566. bottom: "BatchNorm20"
  1567. top: "BatchNorm20"
  1568. scale_param {
  1569. filler {
  1570. value: 1.0
  1571. }
  1572. bias_term: true
  1573. bias_filler {
  1574. value: 0.0
  1575. }
  1576. }
  1577. }
  1578. layer {
  1579. name: "ReLU20"
  1580. type: "ReLU"
  1581. bottom: "BatchNorm20"
  1582. top: "BatchNorm20"
  1583. }
  1584. layer {
  1585. name: "Convolution21"
  1586. type: "Convolution"
  1587. bottom: "BatchNorm20"
  1588. top: "Convolution21"
  1589. convolution_param {
  1590. num_output: 12
  1591. bias_term: false
  1592. pad: 1
  1593. kernel_size: 3
  1594. stride: 1
  1595. weight_filler {
  1596. type: "msra"
  1597. }
  1598. bias_filler {
  1599. type: "constant"
  1600. }
  1601. }
  1602. }
  1603. layer {
  1604. name: "Dropout20"
  1605. type: "Dropout"
  1606. bottom: "Convolution21"
  1607. top: "Dropout20"
  1608. dropout_param {
  1609. dropout_ratio: 0.20000000298
  1610. }
  1611. }
  1612. layer {
  1613. name: "Concat19"
  1614. type: "Concat"
  1615. bottom: "Concat18"
  1616. bottom: "Dropout20"
  1617. top: "Concat19"
  1618. concat_param {
  1619. axis: 1
  1620. }
  1621. }
  1622. layer {
  1623. name: "BatchNorm21"
  1624. type: "BatchNorm"
  1625. bottom: "Concat19"
  1626. top: "BatchNorm21"
  1627. param {
  1628. lr_mult: 0.0
  1629. decay_mult: 0.0
  1630. }
  1631. param {
  1632. lr_mult: 0.0
  1633. decay_mult: 0.0
  1634. }
  1635. param {
  1636. lr_mult: 0.0
  1637. decay_mult: 0.0
  1638. }
  1639. }
  1640. layer {
  1641. name: "Scale21"
  1642. type: "Scale"
  1643. bottom: "BatchNorm21"
  1644. top: "BatchNorm21"
  1645. scale_param {
  1646. filler {
  1647. value: 1.0
  1648. }
  1649. bias_term: true
  1650. bias_filler {
  1651. value: 0.0
  1652. }
  1653. }
  1654. }
  1655. layer {
  1656. name: "ReLU21"
  1657. type: "ReLU"
  1658. bottom: "BatchNorm21"
  1659. top: "BatchNorm21"
  1660. }
  1661. layer {
  1662. name: "Convolution22"
  1663. type: "Convolution"
  1664. bottom: "BatchNorm21"
  1665. top: "Convolution22"
  1666. convolution_param {
  1667. num_output: 12
  1668. bias_term: false
  1669. pad: 1
  1670. kernel_size: 3
  1671. stride: 1
  1672. weight_filler {
  1673. type: "msra"
  1674. }
  1675. bias_filler {
  1676. type: "constant"
  1677. }
  1678. }
  1679. }
  1680. layer {
  1681. name: "Dropout21"
  1682. type: "Dropout"
  1683. bottom: "Convolution22"
  1684. top: "Dropout21"
  1685. dropout_param {
  1686. dropout_ratio: 0.20000000298
  1687. }
  1688. }
  1689. layer {
  1690. name: "Concat20"
  1691. type: "Concat"
  1692. bottom: "Concat19"
  1693. bottom: "Dropout21"
  1694. top: "Concat20"
  1695. concat_param {
  1696. axis: 1
  1697. }
  1698. }
  1699. layer {
  1700. name: "BatchNorm22"
  1701. type: "BatchNorm"
  1702. bottom: "Concat20"
  1703. top: "BatchNorm22"
  1704. param {
  1705. lr_mult: 0.0
  1706. decay_mult: 0.0
  1707. }
  1708. param {
  1709. lr_mult: 0.0
  1710. decay_mult: 0.0
  1711. }
  1712. param {
  1713. lr_mult: 0.0
  1714. decay_mult: 0.0
  1715. }
  1716. }
  1717. layer {
  1718. name: "Scale22"
  1719. type: "Scale"
  1720. bottom: "BatchNorm22"
  1721. top: "BatchNorm22"
  1722. scale_param {
  1723. filler {
  1724. value: 1.0
  1725. }
  1726. bias_term: true
  1727. bias_filler {
  1728. value: 0.0
  1729. }
  1730. }
  1731. }
  1732. layer {
  1733. name: "ReLU22"
  1734. type: "ReLU"
  1735. bottom: "BatchNorm22"
  1736. top: "BatchNorm22"
  1737. }
  1738. layer {
  1739. name: "Convolution23"
  1740. type: "Convolution"
  1741. bottom: "BatchNorm22"
  1742. top: "Convolution23"
  1743. convolution_param {
  1744. num_output: 12
  1745. bias_term: false
  1746. pad: 1
  1747. kernel_size: 3
  1748. stride: 1
  1749. weight_filler {
  1750. type: "msra"
  1751. }
  1752. bias_filler {
  1753. type: "constant"
  1754. }
  1755. }
  1756. }
  1757. layer {
  1758. name: "Dropout22"
  1759. type: "Dropout"
  1760. bottom: "Convolution23"
  1761. top: "Dropout22"
  1762. dropout_param {
  1763. dropout_ratio: 0.20000000298
  1764. }
  1765. }
  1766. layer {
  1767. name: "Concat21"
  1768. type: "Concat"
  1769. bottom: "Concat20"
  1770. bottom: "Dropout22"
  1771. top: "Concat21"
  1772. concat_param {
  1773. axis: 1
  1774. }
  1775. }
  1776. layer {
  1777. name: "BatchNorm23"
  1778. type: "BatchNorm"
  1779. bottom: "Concat21"
  1780. top: "BatchNorm23"
  1781. param {
  1782. lr_mult: 0.0
  1783. decay_mult: 0.0
  1784. }
  1785. param {
  1786. lr_mult: 0.0
  1787. decay_mult: 0.0
  1788. }
  1789. param {
  1790. lr_mult: 0.0
  1791. decay_mult: 0.0
  1792. }
  1793. }
  1794. layer {
  1795. name: "Scale23"
  1796. type: "Scale"
  1797. bottom: "BatchNorm23"
  1798. top: "BatchNorm23"
  1799. scale_param {
  1800. filler {
  1801. value: 1.0
  1802. }
  1803. bias_term: true
  1804. bias_filler {
  1805. value: 0.0
  1806. }
  1807. }
  1808. }
  1809. layer {
  1810. name: "ReLU23"
  1811. type: "ReLU"
  1812. bottom: "BatchNorm23"
  1813. top: "BatchNorm23"
  1814. }
  1815. layer {
  1816. name: "Convolution24"
  1817. type: "Convolution"
  1818. bottom: "BatchNorm23"
  1819. top: "Convolution24"
  1820. convolution_param {
  1821. num_output: 12
  1822. bias_term: false
  1823. pad: 1
  1824. kernel_size: 3
  1825. stride: 1
  1826. weight_filler {
  1827. type: "msra"
  1828. }
  1829. bias_filler {
  1830. type: "constant"
  1831. }
  1832. }
  1833. }
  1834. layer {
  1835. name: "Dropout23"
  1836. type: "Dropout"
  1837. bottom: "Convolution24"
  1838. top: "Dropout23"
  1839. dropout_param {
  1840. dropout_ratio: 0.20000000298
  1841. }
  1842. }
  1843. layer {
  1844. name: "Concat22"
  1845. type: "Concat"
  1846. bottom: "Concat21"
  1847. bottom: "Dropout23"
  1848. top: "Concat22"
  1849. concat_param {
  1850. axis: 1
  1851. }
  1852. }
  1853. layer {
  1854. name: "BatchNorm24"
  1855. type: "BatchNorm"
  1856. bottom: "Concat22"
  1857. top: "BatchNorm24"
  1858. param {
  1859. lr_mult: 0.0
  1860. decay_mult: 0.0
  1861. }
  1862. param {
  1863. lr_mult: 0.0
  1864. decay_mult: 0.0
  1865. }
  1866. param {
  1867. lr_mult: 0.0
  1868. decay_mult: 0.0
  1869. }
  1870. }
  1871. layer {
  1872. name: "Scale24"
  1873. type: "Scale"
  1874. bottom: "BatchNorm24"
  1875. top: "BatchNorm24"
  1876. scale_param {
  1877. filler {
  1878. value: 1.0
  1879. }
  1880. bias_term: true
  1881. bias_filler {
  1882. value: 0.0
  1883. }
  1884. }
  1885. }
  1886. layer {
  1887. name: "ReLU24"
  1888. type: "ReLU"
  1889. bottom: "BatchNorm24"
  1890. top: "BatchNorm24"
  1891. }
  1892. layer {
  1893. name: "Convolution25"
  1894. type: "Convolution"
  1895. bottom: "BatchNorm24"
  1896. top: "Convolution25"
  1897. convolution_param {
  1898. num_output: 12
  1899. bias_term: false
  1900. pad: 1
  1901. kernel_size: 3
  1902. stride: 1
  1903. weight_filler {
  1904. type: "msra"
  1905. }
  1906. bias_filler {
  1907. type: "constant"
  1908. }
  1909. }
  1910. }
  1911. layer {
  1912. name: "Dropout24"
  1913. type: "Dropout"
  1914. bottom: "Convolution25"
  1915. top: "Dropout24"
  1916. dropout_param {
  1917. dropout_ratio: 0.20000000298
  1918. }
  1919. }
  1920. layer {
  1921. name: "Concat23"
  1922. type: "Concat"
  1923. bottom: "Concat22"
  1924. bottom: "Dropout24"
  1925. top: "Concat23"
  1926. concat_param {
  1927. axis: 1
  1928. }
  1929. }
  1930. layer {
  1931. name: "BatchNorm25"
  1932. type: "BatchNorm"
  1933. bottom: "Concat23"
  1934. top: "BatchNorm25"
  1935. param {
  1936. lr_mult: 0.0
  1937. decay_mult: 0.0
  1938. }
  1939. param {
  1940. lr_mult: 0.0
  1941. decay_mult: 0.0
  1942. }
  1943. param {
  1944. lr_mult: 0.0
  1945. decay_mult: 0.0
  1946. }
  1947. }
  1948. layer {
  1949. name: "Scale25"
  1950. type: "Scale"
  1951. bottom: "BatchNorm25"
  1952. top: "BatchNorm25"
  1953. scale_param {
  1954. filler {
  1955. value: 1.0
  1956. }
  1957. bias_term: true
  1958. bias_filler {
  1959. value: 0.0
  1960. }
  1961. }
  1962. }
  1963. layer {
  1964. name: "ReLU25"
  1965. type: "ReLU"
  1966. bottom: "BatchNorm25"
  1967. top: "BatchNorm25"
  1968. }
  1969. layer {
  1970. name: "Convolution26"
  1971. type: "Convolution"
  1972. bottom: "BatchNorm25"
  1973. top: "Convolution26"
  1974. convolution_param {
  1975. num_output: 12
  1976. bias_term: false
  1977. pad: 1
  1978. kernel_size: 3
  1979. stride: 1
  1980. weight_filler {
  1981. type: "msra"
  1982. }
  1983. bias_filler {
  1984. type: "constant"
  1985. }
  1986. }
  1987. }
  1988. layer {
  1989. name: "Dropout25"
  1990. type: "Dropout"
  1991. bottom: "Convolution26"
  1992. top: "Dropout25"
  1993. dropout_param {
  1994. dropout_ratio: 0.20000000298
  1995. }
  1996. }
  1997. layer {
  1998. name: "Concat24"
  1999. type: "Concat"
  2000. bottom: "Concat23"
  2001. bottom: "Dropout25"
  2002. top: "Concat24"
  2003. concat_param {
  2004. axis: 1
  2005. }
  2006. }
  2007. layer {
  2008. name: "BatchNorm26"
  2009. type: "BatchNorm"
  2010. bottom: "Concat24"
  2011. top: "BatchNorm26"
  2012. param {
  2013. lr_mult: 0.0
  2014. decay_mult: 0.0
  2015. }
  2016. param {
  2017. lr_mult: 0.0
  2018. decay_mult: 0.0
  2019. }
  2020. param {
  2021. lr_mult: 0.0
  2022. decay_mult: 0.0
  2023. }
  2024. }
  2025. layer {
  2026. name: "Scale26"
  2027. type: "Scale"
  2028. bottom: "BatchNorm26"
  2029. top: "BatchNorm26"
  2030. scale_param {
  2031. filler {
  2032. value: 1.0
  2033. }
  2034. bias_term: true
  2035. bias_filler {
  2036. value: 0.0
  2037. }
  2038. }
  2039. }
  2040. layer {
  2041. name: "ReLU26"
  2042. type: "ReLU"
  2043. bottom: "BatchNorm26"
  2044. top: "BatchNorm26"
  2045. }
  2046. layer {
  2047. name: "Convolution27"
  2048. type: "Convolution"
  2049. bottom: "BatchNorm26"
  2050. top: "Convolution27"
  2051. convolution_param {
  2052. num_output: 304
  2053. bias_term: false
  2054. pad: 0
  2055. kernel_size: 1
  2056. stride: 1
  2057. weight_filler {
  2058. type: "msra"
  2059. }
  2060. bias_filler {
  2061. type: "constant"
  2062. }
  2063. }
  2064. }
  2065. layer {
  2066. name: "Dropout26"
  2067. type: "Dropout"
  2068. bottom: "Convolution27"
  2069. top: "Dropout26"
  2070. dropout_param {
  2071. dropout_ratio: 0.20000000298
  2072. }
  2073. }
  2074. layer {
  2075. name: "Pooling2"
  2076. type: "Pooling"
  2077. bottom: "Dropout26"
  2078. top: "Pooling2"
  2079. pooling_param {
  2080. pool: AVE
  2081. kernel_size: 2
  2082. stride: 2
  2083. }
  2084. }
  2085. layer {
  2086. name: "BatchNorm27"
  2087. type: "BatchNorm"
  2088. bottom: "Pooling2"
  2089. top: "BatchNorm27"
  2090. param {
  2091. lr_mult: 0.0
  2092. decay_mult: 0.0
  2093. }
  2094. param {
  2095. lr_mult: 0.0
  2096. decay_mult: 0.0
  2097. }
  2098. param {
  2099. lr_mult: 0.0
  2100. decay_mult: 0.0
  2101. }
  2102. }
  2103. layer {
  2104. name: "Scale27"
  2105. type: "Scale"
  2106. bottom: "BatchNorm27"
  2107. top: "BatchNorm27"
  2108. scale_param {
  2109. filler {
  2110. value: 1.0
  2111. }
  2112. bias_term: true
  2113. bias_filler {
  2114. value: 0.0
  2115. }
  2116. }
  2117. }
  2118. layer {
  2119. name: "ReLU27"
  2120. type: "ReLU"
  2121. bottom: "BatchNorm27"
  2122. top: "BatchNorm27"
  2123. }
  2124. layer {
  2125. name: "Convolution28"
  2126. type: "Convolution"
  2127. bottom: "BatchNorm27"
  2128. top: "Convolution28"
  2129. convolution_param {
  2130. num_output: 12
  2131. bias_term: false
  2132. pad: 1
  2133. kernel_size: 3
  2134. stride: 1
  2135. weight_filler {
  2136. type: "msra"
  2137. }
  2138. bias_filler {
  2139. type: "constant"
  2140. }
  2141. }
  2142. }
  2143. layer {
  2144. name: "Dropout27"
  2145. type: "Dropout"
  2146. bottom: "Convolution28"
  2147. top: "Dropout27"
  2148. dropout_param {
  2149. dropout_ratio: 0.20000000298
  2150. }
  2151. }
  2152. layer {
  2153. name: "Concat25"
  2154. type: "Concat"
  2155. bottom: "Pooling2"
  2156. bottom: "Dropout27"
  2157. top: "Concat25"
  2158. concat_param {
  2159. axis: 1
  2160. }
  2161. }
  2162. layer {
  2163. name: "BatchNorm28"
  2164. type: "BatchNorm"
  2165. bottom: "Concat25"
  2166. top: "BatchNorm28"
  2167. param {
  2168. lr_mult: 0.0
  2169. decay_mult: 0.0
  2170. }
  2171. param {
  2172. lr_mult: 0.0
  2173. decay_mult: 0.0
  2174. }
  2175. param {
  2176. lr_mult: 0.0
  2177. decay_mult: 0.0
  2178. }
  2179. }
  2180. layer {
  2181. name: "Scale28"
  2182. type: "Scale"
  2183. bottom: "BatchNorm28"
  2184. top: "BatchNorm28"
  2185. scale_param {
  2186. filler {
  2187. value: 1.0
  2188. }
  2189. bias_term: true
  2190. bias_filler {
  2191. value: 0.0
  2192. }
  2193. }
  2194. }
  2195. layer {
  2196. name: "ReLU28"
  2197. type: "ReLU"
  2198. bottom: "BatchNorm28"
  2199. top: "BatchNorm28"
  2200. }
  2201. layer {
  2202. name: "Convolution29"
  2203. type: "Convolution"
  2204. bottom: "BatchNorm28"
  2205. top: "Convolution29"
  2206. convolution_param {
  2207. num_output: 12
  2208. bias_term: false
  2209. pad: 1
  2210. kernel_size: 3
  2211. stride: 1
  2212. weight_filler {
  2213. type: "msra"
  2214. }
  2215. bias_filler {
  2216. type: "constant"
  2217. }
  2218. }
  2219. }
  2220. layer {
  2221. name: "Dropout28"
  2222. type: "Dropout"
  2223. bottom: "Convolution29"
  2224. top: "Dropout28"
  2225. dropout_param {
  2226. dropout_ratio: 0.20000000298
  2227. }
  2228. }
  2229. layer {
  2230. name: "Concat26"
  2231. type: "Concat"
  2232. bottom: "Concat25"
  2233. bottom: "Dropout28"
  2234. top: "Concat26"
  2235. concat_param {
  2236. axis: 1
  2237. }
  2238. }
  2239. layer {
  2240. name: "BatchNorm29"
  2241. type: "BatchNorm"
  2242. bottom: "Concat26"
  2243. top: "BatchNorm29"
  2244. param {
  2245. lr_mult: 0.0
  2246. decay_mult: 0.0
  2247. }
  2248. param {
  2249. lr_mult: 0.0
  2250. decay_mult: 0.0
  2251. }
  2252. param {
  2253. lr_mult: 0.0
  2254. decay_mult: 0.0
  2255. }
  2256. }
  2257. layer {
  2258. name: "Scale29"
  2259. type: "Scale"
  2260. bottom: "BatchNorm29"
  2261. top: "BatchNorm29"
  2262. scale_param {
  2263. filler {
  2264. value: 1.0
  2265. }
  2266. bias_term: true
  2267. bias_filler {
  2268. value: 0.0
  2269. }
  2270. }
  2271. }
  2272. layer {
  2273. name: "ReLU29"
  2274. type: "ReLU"
  2275. bottom: "BatchNorm29"
  2276. top: "BatchNorm29"
  2277. }
  2278. layer {
  2279. name: "Convolution30"
  2280. type: "Convolution"
  2281. bottom: "BatchNorm29"
  2282. top: "Convolution30"
  2283. convolution_param {
  2284. num_output: 12
  2285. bias_term: false
  2286. pad: 1
  2287. kernel_size: 3
  2288. stride: 1
  2289. weight_filler {
  2290. type: "msra"
  2291. }
  2292. bias_filler {
  2293. type: "constant"
  2294. }
  2295. }
  2296. }
  2297. layer {
  2298. name: "Dropout29"
  2299. type: "Dropout"
  2300. bottom: "Convolution30"
  2301. top: "Dropout29"
  2302. dropout_param {
  2303. dropout_ratio: 0.20000000298
  2304. }
  2305. }
  2306. layer {
  2307. name: "Concat27"
  2308. type: "Concat"
  2309. bottom: "Concat26"
  2310. bottom: "Dropout29"
  2311. top: "Concat27"
  2312. concat_param {
  2313. axis: 1
  2314. }
  2315. }
  2316. layer {
  2317. name: "BatchNorm30"
  2318. type: "BatchNorm"
  2319. bottom: "Concat27"
  2320. top: "BatchNorm30"
  2321. param {
  2322. lr_mult: 0.0
  2323. decay_mult: 0.0
  2324. }
  2325. param {
  2326. lr_mult: 0.0
  2327. decay_mult: 0.0
  2328. }
  2329. param {
  2330. lr_mult: 0.0
  2331. decay_mult: 0.0
  2332. }
  2333. }
  2334. layer {
  2335. name: "Scale30"
  2336. type: "Scale"
  2337. bottom: "BatchNorm30"
  2338. top: "BatchNorm30"
  2339. scale_param {
  2340. filler {
  2341. value: 1.0
  2342. }
  2343. bias_term: true
  2344. bias_filler {
  2345. value: 0.0
  2346. }
  2347. }
  2348. }
  2349. layer {
  2350. name: "ReLU30"
  2351. type: "ReLU"
  2352. bottom: "BatchNorm30"
  2353. top: "BatchNorm30"
  2354. }
  2355. layer {
  2356. name: "Convolution31"
  2357. type: "Convolution"
  2358. bottom: "BatchNorm30"
  2359. top: "Convolution31"
  2360. convolution_param {
  2361. num_output: 12
  2362. bias_term: false
  2363. pad: 1
  2364. kernel_size: 3
  2365. stride: 1
  2366. weight_filler {
  2367. type: "msra"
  2368. }
  2369. bias_filler {
  2370. type: "constant"
  2371. }
  2372. }
  2373. }
  2374. layer {
  2375. name: "Dropout30"
  2376. type: "Dropout"
  2377. bottom: "Convolution31"
  2378. top: "Dropout30"
  2379. dropout_param {
  2380. dropout_ratio: 0.20000000298
  2381. }
  2382. }
  2383. layer {
  2384. name: "Concat28"
  2385. type: "Concat"
  2386. bottom: "Concat27"
  2387. bottom: "Dropout30"
  2388. top: "Concat28"
  2389. concat_param {
  2390. axis: 1
  2391. }
  2392. }
  2393. layer {
  2394. name: "BatchNorm31"
  2395. type: "BatchNorm"
  2396. bottom: "Concat28"
  2397. top: "BatchNorm31"
  2398. param {
  2399. lr_mult: 0.0
  2400. decay_mult: 0.0
  2401. }
  2402. param {
  2403. lr_mult: 0.0
  2404. decay_mult: 0.0
  2405. }
  2406. param {
  2407. lr_mult: 0.0
  2408. decay_mult: 0.0
  2409. }
  2410. }
  2411. layer {
  2412. name: "Scale31"
  2413. type: "Scale"
  2414. bottom: "BatchNorm31"
  2415. top: "BatchNorm31"
  2416. scale_param {
  2417. filler {
  2418. value: 1.0
  2419. }
  2420. bias_term: true
  2421. bias_filler {
  2422. value: 0.0
  2423. }
  2424. }
  2425. }
  2426. layer {
  2427. name: "ReLU31"
  2428. type: "ReLU"
  2429. bottom: "BatchNorm31"
  2430. top: "BatchNorm31"
  2431. }
  2432. layer {
  2433. name: "Convolution32"
  2434. type: "Convolution"
  2435. bottom: "BatchNorm31"
  2436. top: "Convolution32"
  2437. convolution_param {
  2438. num_output: 12
  2439. bias_term: false
  2440. pad: 1
  2441. kernel_size: 3
  2442. stride: 1
  2443. weight_filler {
  2444. type: "msra"
  2445. }
  2446. bias_filler {
  2447. type: "constant"
  2448. }
  2449. }
  2450. }
  2451. layer {
  2452. name: "Dropout31"
  2453. type: "Dropout"
  2454. bottom: "Convolution32"
  2455. top: "Dropout31"
  2456. dropout_param {
  2457. dropout_ratio: 0.20000000298
  2458. }
  2459. }
  2460. layer {
  2461. name: "Concat29"
  2462. type: "Concat"
  2463. bottom: "Concat28"
  2464. bottom: "Dropout31"
  2465. top: "Concat29"
  2466. concat_param {
  2467. axis: 1
  2468. }
  2469. }
  2470. layer {
  2471. name: "BatchNorm32"
  2472. type: "BatchNorm"
  2473. bottom: "Concat29"
  2474. top: "BatchNorm32"
  2475. param {
  2476. lr_mult: 0.0
  2477. decay_mult: 0.0
  2478. }
  2479. param {
  2480. lr_mult: 0.0
  2481. decay_mult: 0.0
  2482. }
  2483. param {
  2484. lr_mult: 0.0
  2485. decay_mult: 0.0
  2486. }
  2487. }
  2488. layer {
  2489. name: "Scale32"
  2490. type: "Scale"
  2491. bottom: "BatchNorm32"
  2492. top: "BatchNorm32"
  2493. scale_param {
  2494. filler {
  2495. value: 1.0
  2496. }
  2497. bias_term: true
  2498. bias_filler {
  2499. value: 0.0
  2500. }
  2501. }
  2502. }
  2503. layer {
  2504. name: "ReLU32"
  2505. type: "ReLU"
  2506. bottom: "BatchNorm32"
  2507. top: "BatchNorm32"
  2508. }
  2509. layer {
  2510. name: "Convolution33"
  2511. type: "Convolution"
  2512. bottom: "BatchNorm32"
  2513. top: "Convolution33"
  2514. convolution_param {
  2515. num_output: 12
  2516. bias_term: false
  2517. pad: 1
  2518. kernel_size: 3
  2519. stride: 1
  2520. weight_filler {
  2521. type: "msra"
  2522. }
  2523. bias_filler {
  2524. type: "constant"
  2525. }
  2526. }
  2527. }
  2528. layer {
  2529. name: "Dropout32"
  2530. type: "Dropout"
  2531. bottom: "Convolution33"
  2532. top: "Dropout32"
  2533. dropout_param {
  2534. dropout_ratio: 0.20000000298
  2535. }
  2536. }
  2537. layer {
  2538. name: "Concat30"
  2539. type: "Concat"
  2540. bottom: "Concat29"
  2541. bottom: "Dropout32"
  2542. top: "Concat30"
  2543. concat_param {
  2544. axis: 1
  2545. }
  2546. }
  2547. layer {
  2548. name: "BatchNorm33"
  2549. type: "BatchNorm"
  2550. bottom: "Concat30"
  2551. top: "BatchNorm33"
  2552. param {
  2553. lr_mult: 0.0
  2554. decay_mult: 0.0
  2555. }
  2556. param {
  2557. lr_mult: 0.0
  2558. decay_mult: 0.0
  2559. }
  2560. param {
  2561. lr_mult: 0.0
  2562. decay_mult: 0.0
  2563. }
  2564. }
  2565. layer {
  2566. name: "Scale33"
  2567. type: "Scale"
  2568. bottom: "BatchNorm33"
  2569. top: "BatchNorm33"
  2570. scale_param {
  2571. filler {
  2572. value: 1.0
  2573. }
  2574. bias_term: true
  2575. bias_filler {
  2576. value: 0.0
  2577. }
  2578. }
  2579. }
  2580. layer {
  2581. name: "ReLU33"
  2582. type: "ReLU"
  2583. bottom: "BatchNorm33"
  2584. top: "BatchNorm33"
  2585. }
  2586. layer {
  2587. name: "Convolution34"
  2588. type: "Convolution"
  2589. bottom: "BatchNorm33"
  2590. top: "Convolution34"
  2591. convolution_param {
  2592. num_output: 12
  2593. bias_term: false
  2594. pad: 1
  2595. kernel_size: 3
  2596. stride: 1
  2597. weight_filler {
  2598. type: "msra"
  2599. }
  2600. bias_filler {
  2601. type: "constant"
  2602. }
  2603. }
  2604. }
  2605. layer {
  2606. name: "Dropout33"
  2607. type: "Dropout"
  2608. bottom: "Convolution34"
  2609. top: "Dropout33"
  2610. dropout_param {
  2611. dropout_ratio: 0.20000000298
  2612. }
  2613. }
  2614. layer {
  2615. name: "Concat31"
  2616. type: "Concat"
  2617. bottom: "Concat30"
  2618. bottom: "Dropout33"
  2619. top: "Concat31"
  2620. concat_param {
  2621. axis: 1
  2622. }
  2623. }
  2624. layer {
  2625. name: "BatchNorm34"
  2626. type: "BatchNorm"
  2627. bottom: "Concat31"
  2628. top: "BatchNorm34"
  2629. param {
  2630. lr_mult: 0.0
  2631. decay_mult: 0.0
  2632. }
  2633. param {
  2634. lr_mult: 0.0
  2635. decay_mult: 0.0
  2636. }
  2637. param {
  2638. lr_mult: 0.0
  2639. decay_mult: 0.0
  2640. }
  2641. }
  2642. layer {
  2643. name: "Scale34"
  2644. type: "Scale"
  2645. bottom: "BatchNorm34"
  2646. top: "BatchNorm34"
  2647. scale_param {
  2648. filler {
  2649. value: 1.0
  2650. }
  2651. bias_term: true
  2652. bias_filler {
  2653. value: 0.0
  2654. }
  2655. }
  2656. }
  2657. layer {
  2658. name: "ReLU34"
  2659. type: "ReLU"
  2660. bottom: "BatchNorm34"
  2661. top: "BatchNorm34"
  2662. }
  2663. layer {
  2664. name: "Convolution35"
  2665. type: "Convolution"
  2666. bottom: "BatchNorm34"
  2667. top: "Convolution35"
  2668. convolution_param {
  2669. num_output: 12
  2670. bias_term: false
  2671. pad: 1
  2672. kernel_size: 3
  2673. stride: 1
  2674. weight_filler {
  2675. type: "msra"
  2676. }
  2677. bias_filler {
  2678. type: "constant"
  2679. }
  2680. }
  2681. }
  2682. layer {
  2683. name: "Dropout34"
  2684. type: "Dropout"
  2685. bottom: "Convolution35"
  2686. top: "Dropout34"
  2687. dropout_param {
  2688. dropout_ratio: 0.20000000298
  2689. }
  2690. }
  2691. layer {
  2692. name: "Concat32"
  2693. type: "Concat"
  2694. bottom: "Concat31"
  2695. bottom: "Dropout34"
  2696. top: "Concat32"
  2697. concat_param {
  2698. axis: 1
  2699. }
  2700. }
  2701. layer {
  2702. name: "BatchNorm35"
  2703. type: "BatchNorm"
  2704. bottom: "Concat32"
  2705. top: "BatchNorm35"
  2706. param {
  2707. lr_mult: 0.0
  2708. decay_mult: 0.0
  2709. }
  2710. param {
  2711. lr_mult: 0.0
  2712. decay_mult: 0.0
  2713. }
  2714. param {
  2715. lr_mult: 0.0
  2716. decay_mult: 0.0
  2717. }
  2718. }
  2719. layer {
  2720. name: "Scale35"
  2721. type: "Scale"
  2722. bottom: "BatchNorm35"
  2723. top: "BatchNorm35"
  2724. scale_param {
  2725. filler {
  2726. value: 1.0
  2727. }
  2728. bias_term: true
  2729. bias_filler {
  2730. value: 0.0
  2731. }
  2732. }
  2733. }
  2734. layer {
  2735. name: "ReLU35"
  2736. type: "ReLU"
  2737. bottom: "BatchNorm35"
  2738. top: "BatchNorm35"
  2739. }
  2740. layer {
  2741. name: "Convolution36"
  2742. type: "Convolution"
  2743. bottom: "BatchNorm35"
  2744. top: "Convolution36"
  2745. convolution_param {
  2746. num_output: 12
  2747. bias_term: false
  2748. pad: 1
  2749. kernel_size: 3
  2750. stride: 1
  2751. weight_filler {
  2752. type: "msra"
  2753. }
  2754. bias_filler {
  2755. type: "constant"
  2756. }
  2757. }
  2758. }
  2759. layer {
  2760. name: "Dropout35"
  2761. type: "Dropout"
  2762. bottom: "Convolution36"
  2763. top: "Dropout35"
  2764. dropout_param {
  2765. dropout_ratio: 0.20000000298
  2766. }
  2767. }
  2768. layer {
  2769. name: "Concat33"
  2770. type: "Concat"
  2771. bottom: "Concat32"
  2772. bottom: "Dropout35"
  2773. top: "Concat33"
  2774. concat_param {
  2775. axis: 1
  2776. }
  2777. }
  2778. layer {
  2779. name: "BatchNorm36"
  2780. type: "BatchNorm"
  2781. bottom: "Concat33"
  2782. top: "BatchNorm36"
  2783. param {
  2784. lr_mult: 0.0
  2785. decay_mult: 0.0
  2786. }
  2787. param {
  2788. lr_mult: 0.0
  2789. decay_mult: 0.0
  2790. }
  2791. param {
  2792. lr_mult: 0.0
  2793. decay_mult: 0.0
  2794. }
  2795. }
  2796. layer {
  2797. name: "Scale36"
  2798. type: "Scale"
  2799. bottom: "BatchNorm36"
  2800. top: "BatchNorm36"
  2801. scale_param {
  2802. filler {
  2803. value: 1.0
  2804. }
  2805. bias_term: true
  2806. bias_filler {
  2807. value: 0.0
  2808. }
  2809. }
  2810. }
  2811. layer {
  2812. name: "ReLU36"
  2813. type: "ReLU"
  2814. bottom: "BatchNorm36"
  2815. top: "BatchNorm36"
  2816. }
  2817. layer {
  2818. name: "Convolution37"
  2819. type: "Convolution"
  2820. bottom: "BatchNorm36"
  2821. top: "Convolution37"
  2822. convolution_param {
  2823. num_output: 12
  2824. bias_term: false
  2825. pad: 1
  2826. kernel_size: 3
  2827. stride: 1
  2828. weight_filler {
  2829. type: "msra"
  2830. }
  2831. bias_filler {
  2832. type: "constant"
  2833. }
  2834. }
  2835. }
  2836. layer {
  2837. name: "Dropout36"
  2838. type: "Dropout"
  2839. bottom: "Convolution37"
  2840. top: "Dropout36"
  2841. dropout_param {
  2842. dropout_ratio: 0.20000000298
  2843. }
  2844. }
  2845. layer {
  2846. name: "Concat34"
  2847. type: "Concat"
  2848. bottom: "Concat33"
  2849. bottom: "Dropout36"
  2850. top: "Concat34"
  2851. concat_param {
  2852. axis: 1
  2853. }
  2854. }
  2855. layer {
  2856. name: "BatchNorm37"
  2857. type: "BatchNorm"
  2858. bottom: "Concat34"
  2859. top: "BatchNorm37"
  2860. param {
  2861. lr_mult: 0.0
  2862. decay_mult: 0.0
  2863. }
  2864. param {
  2865. lr_mult: 0.0
  2866. decay_mult: 0.0
  2867. }
  2868. param {
  2869. lr_mult: 0.0
  2870. decay_mult: 0.0
  2871. }
  2872. }
  2873. layer {
  2874. name: "Scale37"
  2875. type: "Scale"
  2876. bottom: "BatchNorm37"
  2877. top: "BatchNorm37"
  2878. scale_param {
  2879. filler {
  2880. value: 1.0
  2881. }
  2882. bias_term: true
  2883. bias_filler {
  2884. value: 0.0
  2885. }
  2886. }
  2887. }
  2888. layer {
  2889. name: "ReLU37"
  2890. type: "ReLU"
  2891. bottom: "BatchNorm37"
  2892. top: "BatchNorm37"
  2893. }
  2894. layer {
  2895. name: "Convolution38"
  2896. type: "Convolution"
  2897. bottom: "BatchNorm37"
  2898. top: "Convolution38"
  2899. convolution_param {
  2900. num_output: 12
  2901. bias_term: false
  2902. pad: 1
  2903. kernel_size: 3
  2904. stride: 1
  2905. weight_filler {
  2906. type: "msra"
  2907. }
  2908. bias_filler {
  2909. type: "constant"
  2910. }
  2911. }
  2912. }
  2913. layer {
  2914. name: "Dropout37"
  2915. type: "Dropout"
  2916. bottom: "Convolution38"
  2917. top: "Dropout37"
  2918. dropout_param {
  2919. dropout_ratio: 0.20000000298
  2920. }
  2921. }
  2922. layer {
  2923. name: "Concat35"
  2924. type: "Concat"
  2925. bottom: "Concat34"
  2926. bottom: "Dropout37"
  2927. top: "Concat35"
  2928. concat_param {
  2929. axis: 1
  2930. }
  2931. }
  2932. layer {
  2933. name: "BatchNorm38"
  2934. type: "BatchNorm"
  2935. bottom: "Concat35"
  2936. top: "BatchNorm38"
  2937. param {
  2938. lr_mult: 0.0
  2939. decay_mult: 0.0
  2940. }
  2941. param {
  2942. lr_mult: 0.0
  2943. decay_mult: 0.0
  2944. }
  2945. param {
  2946. lr_mult: 0.0
  2947. decay_mult: 0.0
  2948. }
  2949. }
  2950. layer {
  2951. name: "Scale38"
  2952. type: "Scale"
  2953. bottom: "BatchNorm38"
  2954. top: "BatchNorm38"
  2955. scale_param {
  2956. filler {
  2957. value: 1.0
  2958. }
  2959. bias_term: true
  2960. bias_filler {
  2961. value: 0.0
  2962. }
  2963. }
  2964. }
  2965. layer {
  2966. name: "ReLU38"
  2967. type: "ReLU"
  2968. bottom: "BatchNorm38"
  2969. top: "BatchNorm38"
  2970. }
  2971. layer {
  2972. name: "Convolution39"
  2973. type: "Convolution"
  2974. bottom: "BatchNorm38"
  2975. top: "Convolution39"
  2976. convolution_param {
  2977. num_output: 12
  2978. bias_term: false
  2979. pad: 1
  2980. kernel_size: 3
  2981. stride: 1
  2982. weight_filler {
  2983. type: "msra"
  2984. }
  2985. bias_filler {
  2986. type: "constant"
  2987. }
  2988. }
  2989. }
  2990. layer {
  2991. name: "Dropout38"
  2992. type: "Dropout"
  2993. bottom: "Convolution39"
  2994. top: "Dropout38"
  2995. dropout_param {
  2996. dropout_ratio: 0.20000000298
  2997. }
  2998. }
  2999. layer {
  3000. name: "Concat36"
  3001. type: "Concat"
  3002. bottom: "Concat35"
  3003. bottom: "Dropout38"
  3004. top: "Concat36"
  3005. concat_param {
  3006. axis: 1
  3007. }
  3008. }
  3009. layer {
  3010. name: "BatchNorm39"
  3011. type: "BatchNorm"
  3012. bottom: "Concat36"
  3013. top: "BatchNorm39"
  3014. param {
  3015. lr_mult: 0.0
  3016. decay_mult: 0.0
  3017. }
  3018. param {
  3019. lr_mult: 0.0
  3020. decay_mult: 0.0
  3021. }
  3022. param {
  3023. lr_mult: 0.0
  3024. decay_mult: 0.0
  3025. }
  3026. }
  3027. layer {
  3028. name: "Scale39"
  3029. type: "Scale"
  3030. bottom: "BatchNorm39"
  3031. top: "BatchNorm39"
  3032. scale_param {
  3033. filler {
  3034. value: 1.0
  3035. }
  3036. bias_term: true
  3037. bias_filler {
  3038. value: 0.0
  3039. }
  3040. }
  3041. }
  3042. layer {
  3043. name: "ReLU39"
  3044. type: "ReLU"
  3045. bottom: "BatchNorm39"
  3046. top: "BatchNorm39"
  3047. }
  3048. layer {
  3049. name: "Pooling3"
  3050. type: "Pooling"
  3051. bottom: "BatchNorm39"
  3052. top: "Pooling3"
  3053. pooling_param {
  3054. pool: AVE
  3055. global_pooling: true
  3056. }
  3057. }
  3058. layer {
  3059. name: "InnerProduct1"
  3060. type: "InnerProduct"
  3061. bottom: "Pooling3"
  3062. top: "InnerProduct1"
  3063. inner_product_param {
  3064. num_output: 10
  3065. bias_term: true
  3066. weight_filler {
  3067. type: "xavier"
  3068. }
  3069. bias_filler {
  3070. type: "constant"
  3071. }
  3072. }
  3073. }
  3074. layer {
  3075. name: "SoftmaxWithLoss1"
  3076. type: "SoftmaxWithLoss"
  3077. bottom: "InnerProduct1"
  3078. bottom: "label"
  3079. top: "SoftmaxWithLoss1"
  3080. }
  3081. layer {
  3082. name: "Accuracy1"
  3083. type: "Accuracy"
  3084. bottom: "InnerProduct1"
  3085. bottom: "label"
  3086. top: "Accuracy1"
  3087. }
Add Comment
Please, Sign In to add comment