1

Usually to show that our results are not by chance we use significant test like t-test. But when we use 10-fold cross validation we learn&test our modals over chunks of dataset. I'm wondering does we need t-test when we have used 10-fold cross validation? To be more precise I mean when we should use 10-fold cross validation?

user3070752
  • 694
  • 4
  • 23

2 Answers2

1

t-test, when applied to the difference of two normally-distributed variables, is a parametric test that estimates the confidence we have in the fact that the two variables have distinct expectations.

10-fold cross validation is a resampling technique that can be used to estimate confidence we can have in a statistics (e.g., but not only, the difference between two random variables), when we do not know the distribution of the statistics in advance (non-parametric settings).

With cross-validation, you do not get nice 'it is significant' results, you simple have a distribution of the estimations of that statistics, of which the variance is an indicator of the confidence we can have.

Cross validation is often used in model selection when building non-parametric regression or classification models.

damienfrancois
  • 52,978
  • 9
  • 96
  • 110
1

T-test is a type of statistical test on your data. Say you are comparing two datasets and you want to know if the two data sets are significantly different from each other. Then you will do a t-test.

Cross validation is more of a technique for evaluation your models. Typically used to ensure that your model is not overfitting.You do this while building your model to ensure you have a model that is able to generalize to future data. There is not any kind of a statistical test being used here. I hope this helps your doubt.

Trinayan Baruah
  • 323
  • 2
  • 10