I'm implementing object recognition with bags of words histograms. The histograms are composed of 200 "words" per image, kmeans from the descriptors. The problem is that for a large dataset, say 5000 images, we suddenly have 200x5000=1,000,000 words in a histogram. This means that every object will be represented by a 1,000,000 length histogram.
This gets too big and cumbersome past some point. Is there someway around this?