Hello I am training a yolo in a kubeflow pipeline, in order to this, I have a set of pictures more than 1GB.
Currently, I download all images from minio to the container with a script and after that I train the model.
I am not sure if is there any best practice about this, because downloading 1GB per each training is a lot.
is there another way to do this and avoiding building a minio scripts to download picture dataset? can I use a shared volume or something like that in order to share files between operators (the idea is to train another model with the same dataset)