I am trying to understand the security implications of serializing a scikit-learn/keras fitted model (using pickle/joblib etc).
Specifically, if I work on data that I don't want to be revealed, would there be anyway for someone to reverse engineer what data a model was fitted on? Or is the data, just a way for the algorithm to update the relevant coefficients/weights for the algorithm? (If I train the model against "This movie is great" and store it as a foo.pkl file, would I also be able to load the foo.pkl and say it was trained on "This movie is great" if all I have access to is the pkl file and not the data)