Advertisement
Guest User

Untitled

a guest
Jun 25th, 2017
60
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 1.22 KB | None | 0 0
  1. from sklearn.datasets import load_iris
  2. from sklearn.model_selection import train_test_split
  3. import matplotlib.pylot as plt
  4. %matplotlib
  5. import keras
  6. from keras.models import Sequential
  7. from keras.layers import Dense, Activation, Dropout
  8. from keras.utils import to_categorical
  9.  
  10.  
  11. data = load_iris()
  12. X_all = data.data
  13. y_all = to_categorical(data.target)
  14.  
  15. X_tr, X_te, y_tr, y_te = train_test_split(X_all, y_all, test_size=0.33, random_state=42)
  16.  
  17. model = Sequential()
  18. model.add(Dense(4, input_dim=4))
  19. model.add(Activation('relu'))
  20. #model.add(Dropout(0.2))
  21. model.add(Dense(3))
  22. model.add(Activation('softmax'))
  23. model.compile(loss='categorical_crossentropy',
  24. optimizer='sgd',
  25. metrics=['accuracy'])
  26. model.summary()
  27.  
  28. hist=model.fit(X_tr, y_tr, epochs = 500, validation_data=(X_te, y_te))
  29.  
  30.  
  31.  
  32. # summarize history for accuracy
  33. plt.plot(hist.history['acc'])
  34. plt.plot(hist.history['val_acc'])
  35. plt.title('model accuracy')
  36. plt.ylabel('accuracy')
  37. plt.xlabel('epoch')
  38. plt.legend(['train', 'test'], loc='upper left')
  39. plt.show()
  40. # summarize history for loss
  41. plt.plot(hist.history['loss'])
  42. plt.plot(hist.history['val_loss'])
  43. plt.title('model loss')
  44. plt.ylabel('loss')
  45. plt.xlabel('epoch')
  46. plt.legend(['train', 'test'], loc='upper left')
  47. plt.show()
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement