I'm using Airflow with kubernetes executor and the KubernetesPodOperator
. I have two jobs:
- A: Retrieve data from some source up to 100MB
- B: Analyze the data from A.
In order to be able to share the data between the jobs, I would like to run them on the same pod, and then A will write the data to a volume, and B will read the data from the volume.
The documentation states:
The Kubernetes executor will create a new pod for every task instance.
Is there any way to achieve this? And if not, what recommended way there is to pass the data between the jobs?