8

I wish to connect to a remote cluster and execute a Spark process. So, from what I have read, this is specified in the SparkConf.

 val conf = new SparkConf()
  .setAppName("MyAppName")
  .setMaster("spark://my_ip:7077")

Where my_ip is the IP address of my cluster. Unfortunately, I get connection refused. So, I am guessing some credentials must be added to connect correctly. How would I specify the credentials? It seems it would be done with .set(key, value), but have no leads on this.

Alessandro La Corte
  • 419
  • 3
  • 6
  • 18

2 Answers2

4

There are two things missing:

By the way, this would work if you use spark-submit to submit a job, programatically it's more complex to achieve it and could only use yarn-client mode which is tricky to setup remotely.

Varun Tulsian
  • 113
  • 1
  • 7
jamborta
  • 5,130
  • 6
  • 35
  • 55
0
  1. In order to launch application in yarn using spark you should use --master yarn for your spark-submit command or setMaster("yarn") in app configuration initialization.
  2. If case to send "spark-submit" command from remote host can be used popuar Java Secure Channel (JSCH) of course environmental parameter should be set on cluster properly
FaigB
  • 2,271
  • 1
  • 13
  • 22