Hi,
I am trying to setup my windows machine as a remote client by installing
Jupyter+toree to use Spark cluster which is on cloud(IBM Bluemix).
I have installed jupyter+toree. I have also extracted Spark 1.6 on my
windows machineā¦
When I run Jupyter, its works with Python3 kernel, however the Toree Scala
Kernel fails to work.
My objective is to make this Toree-Scala kernel to work with the Spark
Cloud instance.
Where do I need to change the configuration?(my assumption is local
spark-default.conf)
Note: My Spark Cloud instance is configured to use Yarn
My local spark-defaul.conf looks like
# Default system properties included when running spark-submit.
# This is useful for setting default environmental settings.
# Example:
# spark.master spark://master:7077
# spark.eventLog.enabled true
# spark.eventLog.dir hdfs://namenode:8021/directory
# spark.serializer
org.apache.spark.serializer.KryoSerializer
# spark.driver.memory 5g
# spark.executor.extraJavaOptions -XX:+PrintGCDetails -Dkey=value
-Dnumbers="one two three"
Spark Cloud instance spark-default.conf looks like :
spark.yarn.applicationMaster.waitTries 10
spark.yarn.submit.file.replication 3
spark.history.kerberos.principal none
spark.yarn.historyServer.address xxxxxxxxxxxxx:18080
spark.yarn.queue default
spark.yarn.scheduler.heartbeat.interval-ms 5000
spark.history.kerberos.keytab none
spark.driver.extraJavaOptions -Diop.version={{iop_full_version}}
spark.eventLog.dir hdfs://xxxxxxxxxxxx/iop/apps/4.1.0.0/spark/logs/
spark.history.ui.port 18080
spark.yarn.preserve.staging.files false
spark.yarn.driver.memoryOverhead 384
spark.yarn.containerLauncherMaxThreads 25
spark.eventLog.enabled true
spark.yarn.max.executor.failures 3
spark.yarn.jar hdfs://xxxxxxxxxxxx:8020/iop/apps/
4.1.0.0/spark/jars/spar
spark.yarn.am.extraJavaOptions -Diop.version={{iop_full_version}}
spark.yarn.executor.memoryOverhead 384
where xxxxxxxxx was the actual IP address.
--
Vinayak Agrawal
"To Strive, To Seek, To Find and Not to Yield!"
~Lord Alfred Tennyson