I'm using titanic dataset so it's pretty balanced (about 60:40) and the GaussianNB model (standard parameters) has accuracy of 0.659. When I plotted F1, precision and recall I discovered the reason for such a low score.
F1, precision and recall of GaussianNB
Then I calculated ROC-AUC and it's 0.84. I've spent hours trying to understand what happened here, but every answer and blog post is mainly about how misleading ROC-AUC is when used for imbalanced data. What makes ROC-AUC so high, even if we clearly see the model isn't doing very well? Is it because of high recall?
The ROC-AUC Score of LogisticRegression: 0.861
The ROC-AUC Score of LinearDiscriminant: 0.859
The ROC-AUC Score of KNeighbors: 0.855
The ROC-AUC Score of SVC: 0.836
The ROC-AUC Score of GaussianProcess: 0.860
The ROC-AUC Score of DecisionTree: 0.785
The ROC-AUC Score of GaussianNB: 0.840
ROC Curve for other models