I am trying to build a simple one layered LSTM neural network using keras that has a Dense()
output layer with one neuron (used for regression).
I know that the training data when I fit()
the model have to be rescaled. Many online examples rescale both the input and the output within the range [0,1], others within the range [-1,1].
What is the right choice for rescaling the input and the output for a simple regression problem like mine and why? I know it has to do with the activation functions used e.g. sigmoid outputs values within [0,1].
I have seen that you can select the activation function and recurrent activation.
How are the activation functions implemented in the context of an LSTM()
in keras (e.g. the input gate has a 'sigmoid')?
Thanks!