I need to classify images into one of 2000 classes.
I am using the Nvidia DIGITS + caffe (GoogLeNet) and provided 10K samples per class (so a whopping 20 million images, ~1Tb data!). But the data prep ("create db") task itself is being estimated to be 102 days and I shudder to think what the actual training time will be if that estimate is correct.
What is the best way to approach this challenge? should I break up dataset into 3-4 models? and use them separately? Use a smaller dataset and risk less accuracy? something else?
Thanks for helping out a newbie.