I recently conducted a small hackathon , not on a platform like kaggle , but only provided the participants with the training data, and the test data without the true labels.
Is there a way in which I can evaluate their submissions?
I recently conducted a small hackathon , not on a platform like kaggle , but only provided the participants with the training data, and the test data without the true labels.
Is there a way in which I can evaluate their submissions?
You split you train data into train, val and test data. You don't have to use this test data anywhere in the training. It will behave similarly to your actual test data. Run evaluations on this dataset.