Running spark 1.4.1 on CentOS 6.7. Have both python 2.7 and python 3.5.1 installed on it with anaconda.
MAde sure that PYSPARK_PYTHON env var is set to python3.5 but when I open pyspark shell and execute a simple rdd transformation, it errors out with below exception:
Exception: Python in worker has different version 2.7 than that in driver 3.5, PySpark cannot run with different minor versions
Just wondering what are the other places to change the path.