Train LSTM model using keras on the given dataset using Glove Embeddings available here - (https://nlp.stanford.edu/projects/glove/)
| data | the sentiment140 train dataset with  | 
|---|---|
| max_words | Maximum number of words to consider using word frequency measure. | 
| maxlen | Maximum length of a sequence. | 
| embedding_dim | Output dimension of the embedding layer. | 
| epochs | Number of epochs to run the training for. | 
| batch_size | Batch Size for model fitting. | 
| validation_split | Split ratio for validation | 
| lstm_units | Number of units i.e. output dimension of lstm layer. | 
| seed | Seed for shuffling training data. | 
| glove_file_path | File path location for glove embeddings. | 
| model_save_path | File path location for saving model. | 
plot of the training operation showing train vs validation loss and accuracy.
# NOT RUN { data(sentiment140_train) train_lstm_with_glove(glove_file_path = "./glove.6B.100d.txt", model_save_path = "./train_glove_lstm.h5") # }