Advertisement
Guest User

Untitled

a guest
Mar 25th, 2017
244
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 2.24 KB | None | 0 0
  1. K.set_image_dim_ordering('th')
  2. model = Sequential()
  3.  
  4.  
  5. #first set of CONV => CONV => CONV => LReLU => MAXPOOL
  6. model.add(Convolution2D(64, kernel_size=(3, 3), padding="same", data_format='channels_first', input_shape = (d, h, w), kernel_initializer = 'glorot_normal', bias_initializer=constant(0.1) ))
  7. model.add(LeakyReLU(alpha=alp))
  8. model.add(Convolution2D(64, kernel_size=(3, 3), border_mode="same", data_format='channels_first', input_shape = (64, 33, 33), kernel_initializer = 'glorot_normal', bias_initializer=constant(0.1) ))
  9. model.add(LeakyReLU(alpha=alp))
  10. model.add(Convolution2D(64, kernel_size=(3, 3), border_mode="same", data_format='channels_first', input_shape = (64, 33, 33), kernel_initializer = 'glorot_normal', bias_initializer=constant(0.1) ))
  11. model.add(LeakyReLU(alpha=alp))
  12. model.add(MaxPooling2D(pool_size=(3, 3), strides=(2, 2)))
  13.  
  14. #second set of CONV => CONV => CONV => LReLU => MAXPOOL
  15. model.add(Convolution2D(128, kernel_size=(3, 3), border_mode="same", data_format='channels_first', input_shape = (64, 16, 16), kernel_initializer = 'glorot_normal', bias_initializer=constant(0.1) ))
  16. model.add(LeakyReLU(alpha=alp))
  17. model.add(Convolution2D(128, kernel_size=(3, 3), border_mode="same", data_format='channels_first', input_shape = (128, 16, 16), kernel_initializer = 'glorot_normal', bias_initializer=constant(0.1) ))
  18. model.add(LeakyReLU(alpha=alp))
  19. model.add(Convolution2D(128, kernel_size=(3, 3), border_mode="same", data_format='channels_first', input_shape = (128, 16, 16), kernel_initializer = 'glorot_normal', bias_initializer=constant(0.1) ))
  20. model.add(LeakyReLU(alpha = alp))
  21. model.add(MaxPooling2D(pool_size=(3, 3), strides=(2, 2)))
  22.  
  23. #Fully connected layers
  24.  
  25. # FC => LReLU => FC => LReLU
  26. model.add(Flatten())
  27. model.add(Dense(256, kernel_initializer = 'glorot_normal', bias_initializer=constant(0.1)))
  28. model.add(LeakyReLU(alp))
  29. model.add(Dropout(dropout))
  30. model.add(Dense(256, kernel_initializer = 'glorot_normal', bias_initializer=constant(0.1)))
  31. model.add(LeakyReLU(alp))
  32. model.add(Dropout(dropout))
  33.  
  34. # FC => SOFTMAX
  35. model.add(Dense(classes, kernel_initializer = 'glorot_normal', bias_initializer = constant(0.1)))
  36. model.add(Activation("softmax"))
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement