I am working on a spark program in which I have to load avro data and process it. I am trying to understand how the job ids are created for a spark application. I use the below line of code to load the avro data.
sqlContext.read.format("com.databricks.spark.avro").load(path)
As far as I know job ids will be created based on the actions encountered in the program. My job is scheduled to run every 30 mins. When I look at the spark history server for this application, I see a job id is created for the load operation. It happens only sometimes and log looks absolutely fine. I am using spark 1.6.1
I am curios to know if the load operation creates new job id in an application?