While running a spark-submit
on a spark standalone cluster comprised of one master and 1 worker, the caffe
python module does not get imported due to error ImportError: No module named caffe
This doesn't seem to be an issue whenever I run a job locally by
spark-submit --master local script.py
the caffe
module gets imported just fine.
The environmental variables are currently set under ~/.profile
for spark and caffe and they are pointing to the PYTHONPATH
.
Is ~/.profile
the correct location to set these variables or perhaps a system wide configuration is needed such as adding the variables under /etc/profile.d/