In order to parallelize the read operation and read with more than one executor. Rather than the following read code, I want to read with JDBC.
hosts ={"spark.cassandra.connection.host":'node1_ip,node2_ip,node3_ip',
"table":"ex_table","keyspace":"ex_keyspace"}
data_frame=sqlContext.read.format("org.apache.spark.sql.cassandra") \
.options(**hosts).load()
How can I read Cassandra data using JDBC from pySpark?