I am facing tough times deploying Dataflow pipelines. Thanks to GCP Documents.. Below is what I am trying to achieve.
I have 4 deep learning models (binary files each 1 gb). I want to get predictions from all the 4 models. So I stired all the 4 models in bucket. And in my pipeline I do.
download_blob(......, destination_file_name = 'model.bin')
fasttext.load_model('model.bin')
It works fine but I have below concern.
Everytime a job is created it downloads these files which will consume lot of time. If I call 100 jobs, So the models will be downloaded 100 times. Is there any way I could avoid it?
Is there any way I could stage these files in some location so that even if I trigger job 100 times the model is downloaded just one time?