I would like to use the spark-csv package with SparkR using RStudio. It works perfectly with the SparkR shell but I didn't find any way to include it in a RStudio session.
Any idea how to do it ?
Thanks for your help
I would like to use the spark-csv package with SparkR using RStudio. It works perfectly with the SparkR shell but I didn't find any way to include it in a RStudio session.
Any idea how to do it ?
Thanks for your help
I have had the same problem, look to this question
The solution given by Pragith works perfect without building the assembly jar: run
Sys.setenv('SPARKR_SUBMIT_ARGS'='"--packages" "com.databricks:spark-csv_2.10:1.0.3" "sparkr-shell"')
before
library(SparkR)
And you can read the .csv file from within RStudio. In the same way, you should be able to include all other packages you want.