Advertisement
Guest User

Untitled

a guest
Jun 19th, 2019
81
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 1.17 KB | None | 0 0
  1. # LSTM for sequence classification in the IMDB dataset
  2. import numpy
  3. from keras.datasets import imdb
  4. from keras.models import Sequential
  5. from keras.layers import Dense
  6. from keras.layers import LSTM
  7. from keras.layers.embeddings import Embedding
  8. from keras.preprocessing import sequence
  9. # fix random seed for reproducibility
  10. numpy.random.seed(7)
  11. # load the dataset but only keep the top n words, zero the rest
  12. top_words = 5000
  13. (X_train, y_train), (X_test, y_test) = imdb.load_data(num_words=top_words)
  14. # truncate and pad input sequences
  15. max_review_length = 500
  16. X_train = sequence.pad_sequences(X_train, maxlen=max_review_length)
  17. X_test = sequence.pad_sequences(X_test, maxlen=max_review_length)
  18. # create the model
  19. embedding_vecor_length = 32
  20. model = Sequential()
  21. model.add(Embedding(top_words, embedding_vecor_length, input_length=max_review_length))
  22. model.add(LSTM(100))
  23. model.add(Dense(1, activation='sigmoid'))
  24. model.compile(loss='binary_crossentropy', optimizer='adam', metrics=['accuracy'])
  25. print(model.summary())
  26. model.fit(X_train, y_train, epochs=3, batch_size=64)
  27. # Final evaluation of the model
  28. scores = model.evaluate(X_test, y_test, verbose=0)
  29. print("Accuracy: %.2f%%" % (scores[1]*100))
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement