I am building a CBIR application . I am using the features extracted from a deep convnet. The feature vectors are quite big ( about 100,000 in size) . And the dataset has more than 10k images. I have already gone through the answer to this problem, and I don't want to use the libraries mentioned in the same.
I tried cPickle and hdf5 for storing the feature vectors. I am running it on a PC having 4GB of RAM and a 2Ghz Intel Core i3 processor.
The following piece of code builds the index :
h = h5py.File(index_file, 'w')
for imagePath in glob.glob(args["dataset"] + "/*.*"):
# extract our unique image ID (i.e. the filename)
k = imagePath[imagePath.rfind('/') + 1:]
features = get_features(imagePath, args["layer"])
h.create_dataset(k, data=features)
Whenever I run the program to build the index for my dataset of images, I get the error "Python.exe has stopped working"after around 16MB of index file has been created. I am new to hdf5 and the answer maybe trivial, but any help would be deeply appreciated.