I'm facing the following issue. I have a large number of documents that I want to encode using a bidirectional LSTM. Each document has a different number of words and word can be thought of as a timestep.
When configuring the bidirectional LSTM we are expected to provide the timeseries length.
When I am training the model this value will be different for each batch.
Should I choose a number for the timeseries_size
which is the biggest document size I will allow? Any documents bigger than this will not be encoded?
Example config:
Bidirectional(LSTM(128, return_sequences=True), input_shape=(timeseries_size, encoding_size))