I have a question about Naïve Bayes classifier with skewed data distribution for training and test data.
- training data has 90% spam and 10% non-spam
- test data has 80% non- spam and 20% spam
Would it be better to use MLE(max. likelihood) than MAP(standard max. posterior probability) for decision function for training data or not?
My understanding is as the distribution of training data and that of test data is different, if we use max. posterior probabilities then test results will be biased towards spam class, So MLE is better. Is my understanding correct?