Advertisement
Guest User

Untitled

a guest
Feb 21st, 2018
52
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 1.95 KB | None | 0 0
  1. model = Sequential()
  2. model.add(BatchNormalization(input_shape=(96,128,3)))
  3.  
  4. # Convolution no.1
  5. model.add(
  6. Conv2D(
  7. filters=16,
  8. kernel_size=(5, 5),
  9. strides=(2, 2),
  10. bias_initializer='he_normal',
  11. padding="valid"))
  12. model.add(LeakyReLU())
  13.  
  14. model.add(MaxPooling2D(pool_size=(2, 2)))
  15. model.add(BatchNormalization())
  16.  
  17. # Convolution no.2
  18. model.add(
  19. Conv2D(
  20. filters=24,
  21. kernel_size=(5, 5),
  22. strides=(2, 2),
  23. bias_initializer='he_normal',
  24. padding="valid"))
  25. model.add(LeakyReLU())
  26. model.add(BatchNormalization())
  27.  
  28. # Convolution no.3
  29. model.add(
  30. Conv2D(
  31. filters=32,
  32. kernel_size=(3, 3),
  33. bias_initializer='he_normal',
  34. padding="valid"))
  35. model.add(LeakyReLU())
  36. model.add(BatchNormalization())
  37.  
  38. model.add(Flatten())
  39.  
  40. # Fully connected no.1
  41. model.add(
  42. Dense(
  43. units=300,
  44. bias_initializer='he_normal',
  45. activation='relu'))
  46. model.add(BatchNormalization())
  47. model.add(Dropout(0.25))
  48.  
  49. # Fully connected no.2
  50. model.add(
  51. Dense(
  52. units=100,
  53. bias_initializer='he_normal',
  54. activation='relu'))
  55. model.add(BatchNormalization())
  56. model.add(Dropout(0.25))
  57.  
  58. # Fully connected no.3
  59. model.add(
  60. Dense(
  61. units=1,
  62. bias_initializer='he_normal',
  63. activation='tanh',
  64. name='prediction'))
  65.  
  66. opt = optimizers.SGD(lr=0.01)
  67. model.compile(loss='mse', optimizer=opt, metrics=['mse'])
  68.  
  69. return model
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement