Good morning,
I trained a LSTM network on yelp https://www.yelp.com/dataset restaurants data set. It is a large dataset and it took several days to train on my PC. Anyways I saved the model and weights and now wish to use it for predictions for real time sentiment evaluations.
What is the common / good / best practice to do this: I load the model and the weights, I then compile it. This is not an issue there are plenty examples in the documentation or on the Internet. However what next? All I need to do is to tokenize the newly received review then pad it and pass to the model.predict?
tokenizer = Tokenizer(num_words = 2500, split=' ')
tokenizer.fit_on_texts(data['text'].values)
print(tokenizer.word_index)
X = tokenizer.texts_to_sequences(data['text'].values)
X = pad_sequences(X)
Cannot be that simple⦠If it is all what is required then how this is connected with the tokenizer that was used to train the model? It was an expensive operation to tokenize more than 2.5 milion reviews downloaded originally from yelp dataset?
Thank you for any suggestions.