Advertisement
Guest User

Untitled

a guest
Jan 13th, 2017
71
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
  1. name: "Hyb2Net-LSTM"
  2. input: "data"
  3. input_dim: 180
  4. input_dim: 3
  5. input_dim: 227
  6. input_dim: 227
  7. input: "clip_markers"
  8. input_dim: 180
  9. input_dim: 1
  10. input_dim: 1
  11. input_dim: 1
  12.  
  13. layer {
  14.   name: "conv1"
  15.   type: "Convolution"
  16.   bottom: "data"
  17.   top: "conv1"
  18.   convolution_param {
  19.     num_output: 96
  20.     kernel_size: 7
  21.     stride: 2
  22.     weight_filler {
  23.       type: "gaussian"
  24.       std: 0.01
  25.     }
  26.     bias_filler {
  27.       type: "constant"
  28.       value: 0.1
  29.     }
  30.   }
  31. }
  32. layer {
  33.   name: "relu1"
  34.   type: "ReLU"
  35.   bottom: "conv1"
  36.   top: "conv1"
  37. }
  38. layer {
  39.   name: "pool1"
  40.   type: "Pooling"
  41.   bottom: "conv1"
  42.   top: "pool1"
  43.   pooling_param {
  44.     pool: MAX
  45.     kernel_size: 3
  46.     stride: 2
  47.   }
  48. }
  49. layer {
  50.   name: "norm1"
  51.   type: "LRN"
  52.   bottom: "pool1"
  53.   top: "norm1"
  54.   lrn_param {
  55.     local_size: 5
  56.     alpha: 0.0001
  57.     beta: 0.75
  58.   }
  59. }
  60. layer {
  61.   name: "conv2"
  62.   type: "Convolution"
  63.   bottom: "norm1"
  64.   top: "conv2"
  65.   convolution_param {
  66.     num_output: 384
  67.     kernel_size: 5
  68.     group: 2
  69.     stride: 2
  70.     weight_filler {
  71.       type: "gaussian"
  72.       std: 0.01
  73.     }
  74.     bias_filler {
  75.       type: "constant"
  76.       value: 0.1
  77.     }
  78.   }
  79. }
  80. layer {
  81.   name: "relu2"
  82.   type: "ReLU"
  83.   bottom: "conv2"
  84.   top: "conv2"
  85. }
  86. layer {
  87.   name: "pool2"
  88.   type: "Pooling"
  89.   bottom: "conv2"
  90.   top: "pool2"
  91.   pooling_param {
  92.     pool: MAX
  93.     kernel_size: 3
  94.     stride: 2
  95.   }
  96. }
  97. layer {
  98.   name: "norm2"
  99.   type: "LRN"
  100.   bottom: "pool2"
  101.   top: "norm2"
  102.   lrn_param {
  103.     local_size: 5
  104.     alpha: 0.0001
  105.     beta: 0.75
  106.   }
  107. }
  108. layer {
  109.   name: "conv3"
  110.   type: "Convolution"
  111.   bottom: "norm2"
  112.   top: "conv3"
  113.   convolution_param {
  114.     num_output: 512
  115.     pad: 1
  116.     kernel_size: 3
  117.     weight_filler {
  118.       type: "gaussian"
  119.       std: 0.01
  120.     }
  121.     bias_filler {
  122.       type: "constant"
  123.       value: 0.1
  124.     }
  125.   }
  126. }
  127. layer {
  128.   name: "relu3"
  129.   type: "ReLU"
  130.   bottom: "conv3"
  131.   top: "conv3"
  132. }
  133. layer {
  134.   name: "conv4"
  135.   type: "Convolution"
  136.   bottom: "conv3"
  137.   top: "conv4"
  138.   convolution_param {
  139.     num_output: 512
  140.     pad: 1
  141.     kernel_size: 3
  142.     group: 2
  143.     weight_filler {
  144.       type: "gaussian"
  145.       std: 0.01
  146.     }
  147.     bias_filler {
  148.       type: "constant"
  149.       value: 0.1
  150.     }
  151.   }
  152. }
  153. layer {
  154.   name: "relu4"
  155.   type: "ReLU"
  156.   bottom: "conv4"
  157.   top: "conv4"
  158. }
  159. layer {
  160.   name: "conv5"
  161.   type: "Convolution"
  162.   bottom: "conv4"
  163.   top: "conv5"
  164.   convolution_param {
  165.     num_output: 384
  166.     pad: 1
  167.     kernel_size: 3
  168.     group: 2
  169.     weight_filler {
  170.       type: "gaussian"
  171.       std: 0.01
  172.     }
  173.     bias_filler {
  174.       type: "constant"
  175.       value: 0.1
  176.     }
  177.   }
  178. }
  179. layer {
  180.   name: "relu5"
  181.   type: "ReLU"
  182.   bottom: "conv5"
  183.   top: "conv5"
  184. }
  185. layer {
  186.   name: "pool5"
  187.   type: "Pooling"
  188.   bottom: "conv5"
  189.   top: "pool5"
  190.   pooling_param {
  191.     pool: MAX
  192.     kernel_size: 3
  193.     stride: 2
  194.   }
  195. }
  196. layer {
  197.   name: "fc6"
  198.   type: "InnerProduct"
  199.   bottom: "pool5"
  200.   top: "fc6"
  201.   param {
  202.     lr_mult: 1
  203.     decay_mult: 1
  204.   }
  205.   param {
  206.     lr_mult: 2
  207.     decay_mult: 0
  208.   }
  209.   inner_product_param {
  210.     num_output: 4096
  211.     weight_filler {
  212.       type: "gaussian"
  213.       std: 0.01
  214.     }
  215.     bias_filler {
  216.       type: "constant"
  217.       value: 0.1
  218.     }
  219.   }
  220. }
  221. layer {
  222.   name: "relu6"
  223.   type: "ReLU"
  224.   bottom: "fc6"
  225.   top: "fc6"
  226. }
  227. layer{
  228.   name: "reshape-data"
  229.   type: "Reshape"
  230.   bottom: "fc6"
  231.   top: "fc6-reshape"
  232.   reshape_param{
  233.     shape{
  234.       dim: 18
  235.       dim: 10
  236.       dim: 4096
  237.     }
  238.   }
  239.   include: { phase: TEST}
  240. }
  241. layer{
  242.   name: "reshape-cm"
  243.   type: "Reshape"
  244.   bottom: "clip_markers"
  245.   top: "reshape-cm"
  246.   reshape_param{
  247.     shape{
  248.       dim: 18
  249.       dim: 10
  250.     }
  251.   }
  252.   include: { phase: TEST }
  253. }
  254. layer {
  255.   name: "lstm1"
  256.   type: "LSTM"
  257.   bottom: "fc6-reshape"
  258.   bottom: "reshape-cm"
  259.   top: "lstm1"
  260.   recurrent_param {
  261.     num_output: 256
  262.     bias_filler {
  263.       type: "constant"
  264.       value: 0
  265.     }
  266.   }
  267. }
  268. layer {
  269.   name: "fc8-final"
  270.   type: "InnerProduct"
  271.   bottom: "lstm1"
  272.   top: "fc8-final"
  273.   param {
  274.     lr_mult: 10
  275.     decay_mult: 1
  276.   }
  277.   param {
  278.     lr_mult: 20
  279.     decay_mult: 0
  280.   }
  281.   inner_product_param {
  282.     num_output: 2
  283.     weight_filler {
  284.       type: "gaussian"
  285.       std: 0.01
  286.     }
  287.     bias_filler {
  288.       type: "constant"
  289.       value: 0
  290.     }
  291.     axis: 2
  292.   }
  293. }
  294. layer {
  295.   name: "probs"
  296.   type: "Softmax"
  297.   bottom: "fc8-final"
  298.   top: "probs"
  299.   softmax_param {
  300.     axis: 2
  301.   }
  302. }
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement