I see papers that use 10-fold cross validation on data sets that have a number of samples indivisible by 10.
I couldn't find any case where they explained how they chose each subset.
My assumption is that they use resampling to some extent, but if this were to be the case then a sample could appear in both subsets and therefore bias the model.
Paper as example: http://www.biomedcentral.com/1471-2105/9/319
Would it be recommended to do the following;
- Given a sample size of 86 take 8 samples as a holdout set.
- Use the remaining samples to train.
- repeat 10 times.
Doing it this way would have it so every sample is a training set but only 80/86 samples are used as holdouts and there is no bias of having it occur within both a training and holdout set.
Any insight would be appreciated.