I am trying to run two spark actions as below and I expect them to run parallely as they both use differenct pools. Does scheduling using pools meant that, different independent actions will run parallelly? I mean If I have 200 cores, then pool1 uses 100 cores and pool2 uses 100 cores and then process the action. In my case after first dataframe action is completed in pool1 then dataframe action2 is started.
spark.setLocalProperty("spark.scheduler.pool","pool1")
dataframe.show(100,false)
spark.setLocalProperty("spark.scheduler.pool","pool2")
dataframe2.show(100,false)
My pool configuration xml
<?xml version="1.0"?>
<allocations>
<pool name="pool1">
<schedulingMode>FAIR</schedulingMode>
<weight>1</weight>
</pool>
<pool name="pool2">
<schedulingMode>FAIR</schedulingMode>
<weight>1</weight>
</pool>
</allocations>