Train LSTM model using keras on the given dataset using Glove Embeddings available here - (https://nlp.stanford.edu/projects/glove/)
data | the sentiment140 train dataset with |
---|---|
max_words | Maximum number of words to consider using word frequency measure. |
maxlen | Maximum length of a sequence. |
embedding_dim | Output dimension of the embedding layer. |
epochs | Number of epochs to run the training for. |
batch_size | Batch Size for model fitting. |
validation_split | Split ratio for validation |
lstm_units | Number of units i.e. output dimension of lstm layer. |
seed | Seed for shuffling training data. |
glove_file_path | File path location for glove embeddings. |
model_save_path | File path location for saving model. |
plot of the training operation showing train vs validation loss and accuracy.
# NOT RUN { data(sentiment140_train) train_lstm_with_glove(glove_file_path = "./glove.6B.100d.txt", model_save_path = "./train_glove_lstm.h5") # }