Any ideas how to read aws s3 with scala. I tried this link
https://www.cloudera.com/documentation/enterprise/5-5-x/topics/spark_s3.html
But could not get it to work. I can do so the same in data bricks but with dsx its not working etc.
IBM has documented steps for python here but none for scala - https://datascience.ibm.com/blog/use-ibm-data-science-experience-to-read-and-write-data-stored-on-amazon-s3/
spark.sparkContext.hadoopConfiguration.set("fs.s3n.impl","org.apache.hadoop.fs.s3native.NativeS3FileSystem") spark.sparkContext.hadoopConfiguration.set("fs.s3n.awsAccessKeyId", "xyz") spark.sparkContext.hadoopConfiguration.set("fs.s3n.awsSecretAccessKey", "abc")
val df_data_1 = spark.read.format("csv").option("header", "true").load("s3a://defg/retail-data/by-day/*.csv") df_data_1.take(5)