I'm trying to create an RStudio-like experience for Scala and spark. After figuring out how to install it properly (including Scala 2.10.5) on Win7x64, I'm facing the problem, that the version of the interpreter
scala> scala.tools.nsc.Properties.versionString
res1: String = version 2.11.8
differs from the spark-shell
scala version
scala> scala.tools.nsc.Properties.versionString
res1: String = version 2.10.5
and from the scala version (cmd.exe
)
C:\>scala -version
Scala code runner version 2.10.5 -- Copyright 2002-2013, LAMP,EPFL
Which causes my problems according to this answer problems when starting spark from the interpreter
scala> val sc = new SparkContext(conf)
java.lang.NoSuchMethodError: scala.
collection.immutable.HashSet$.empty()Lscala/collection/immutable/HashSet;
Question: Where do I set/configure the version the interpreter uses? Searching the global settings of the Scala IDE for "interpreter" gave no results. Seems to be something that was shipped along with the Scala IDE?
(spark 1.6.1, hadoop 2.6, Scala IDE 4.4.1, Scala 2.10.5, at least that's the way it should be)