I want to read parquetFile in sparkR shell from the hdfs system. So I do the that:
./sparkR --master yarn-client
sqlContext <- sparkRSQL.init(sc)
path<-"hdfs://year=2015/month=1/day=9"
AppDF <- parquetFile(sqlContext, path)
Error: No such file or directory
But this file is really exist in the hdfs system. And when I wrap this code in R file like dataframe.R and run ./spark-submit --master yarn ~/dataframe.R 1000
. It works well. So I think the problem is running on yarn-client through sparkR shell. Could anyone help to solve this?
I'm using spark-1.4.0-bin-hadoop2.6