I am using caffe, a deep neural network library, to generate image features for image based retrieval. The particular network I am using generates a 4096 dimensional feature.
I am using LSHash to generate hash buckets from the features. When I do a brute for comparison of all available feature, sorting images by euclidean distance, I find the features represent image similarity well. When I use LSHash, however, I find that similar features rarely land in the same bucket.
Are the source features too large for use with LSH? Are there other ways to reduce the dimensions of the image features before attempting to hash them?