So, I am wondering if there is any difference between the StandardScaler of Spark and a simple z-score calculation.
The formula for the z-score calculation is:
z = (x-mean)/Std
However for the StandardScaler of Spark it is not clear to me how exactly it normalizes values. I could not find any formula for this. Assuming that we set both "setWithStd" and "setWithMean" to true as below:
StandardScaler scaler = new StandardScaler()
.setInputCol("features")
.setOutputCol("scaledFeatures")
.setWithStd(true)
.setWithMean(true);
Would it be the same as using a simple z-score calculation?