I'm experimenting with Google Cloud Data Fusion. I'm joining 2 BigQuery tables using the joiner + writing back to BigQuery. In preview I get this error : java.lang.NoSuchMethodError: net.jpountz.lz4.LZ4BlockInputStream.(Ljava/io/InputStream;Z)V
I've set the job to Spark instead of Map/Reduce because Map/reduce generates an out of memory issue in preview. When I deploy and run the job it crashes due to "Container killed by YARN for exceeding memory limits."
The largest tables is about 6 million records without any nested fields. The smaller table is 66 records.
I didn't specify any partitions.
What's the recommended way to debug/solve this issue? Increase the number of workers / memory?