I'm working on a neural network to predict if an event is a neutrino or not. Each input x is a matrix : (3, 5484), three different measures, 5484 sensors.
So I should normalize the input for the network to work properly but I have doubts about it, let's explain :
My dataset is : (x_1,....,x_N). Let's focus on just one feature and think x is a vector of size (5484). Then each x has a mean : mean_x, and a std :std_x But the point of normalizing input is to rescale it without losing information (global normalization, same normalizing factors for each input) right ?
So, I'm not sure how I should rescale it. Should I flatten (x_1, .., x_N) to a 1-D vector : (x_1, ..., x_5484*N) and compute its mean and standard deviation ?
Or should I treat independently each of the 5484 input channels ?
I'm kind of lost.