I just tried this with using zeppelin master, and the classnotfound error is still there.
On Thu, Jul 7, 2016 at 7:07 AM, Hyung Sung Shim <hss...@nflabs.com> wrote: > Hi. Jonathan. > If it is possible, could you use the zeppelin master on the github? > Zeppelin is working well with spark authentication when i tested. > > > > 2016-07-02 3:05 GMT+09:00 Jonathan Esterhazy <jonathan.esterh...@gmail.com > >: > >> Hyung, thx for your help. I opened these: >> >> https://issues.apache.org/jira/browse/ZEPPELIN-1096 (this scala problem) >> https://issues.apache.org/jira/browse/ZEPPELIN-1097 (similar looking >> python problem) >> >> LMK if I can provide more info or help in some way. >> >> On Fri, Jul 1, 2016 at 5:08 AM, Hyung Sung Shim <hss...@nflabs.com> >> wrote: >> >>> Hi Jonathan. >>> Unfortunately I got same error in my test bed. >>> Do you mind create an jira issue for this? >>> >>> 2016-07-01 3:19 GMT+09:00 Jonathan Esterhazy < >>> jonathan.esterh...@gmail.com>: >>> >>>> I added this and still get the same exception. The same property is >>>> also set in spark-defaults.conf. >>>> >>>> After that didn't work, I also tried adding --conf >>>> spark.authenticate=true and --conf >>>> spark.authenticate.enableSaslEncryption=true, to match the other related >>>> settings in spark-defaults.conf. Still get the same classnotfoundexception. >>>> >>>> >>>> On Thu, Jun 30, 2016 at 10:45 AM, Hyung Sung Shim <hss...@nflabs.com> >>>> wrote: >>>> >>>>> Please add the *export SPARK_SUBMIT_OPTIONS="--conf >>>>> spark.authenticate.secret=secret"* in conf/zeppelin-env.sh and >>>>> restart zeppelin and retry your code. >>>>> >>>>> >>>>> 2016-06-30 23:34 GMT+09:00 Jonathan Esterhazy < >>>>> jonathan.esterh...@gmail.com>: >>>>> >>>>>> yes it does. i only see this problem in zeppelin. >>>>>> >>>>>> On Thu, Jun 30, 2016 at 7:05 AM, Hyung Sung Shim <hss...@nflabs.com> >>>>>> wrote: >>>>>> >>>>>>> Hi Jonathan. >>>>>>> It's not easy to build the test environments but I am working on >>>>>>> this. >>>>>>> I have question for you. >>>>>>> Does your code working well on spark-shell in the spark.authenticate >>>>>>> mode? >>>>>>> >>>>>>> 2016-06-30 22:47 GMT+09:00 Jonathan Esterhazy < >>>>>>> jonathan.esterh...@gmail.com>: >>>>>>> >>>>>>>> Hyung, did you have any luck w/ zeppelin + spark authentication? >>>>>>>> I'm quite stumped. >>>>>>>> >>>>>>>> thx. >>>>>>>> >>>>>>>> On Tue, Jun 28, 2016 at 9:11 PM, Hyung Sung Shim <hss...@nflabs.com >>>>>>>> > wrote: >>>>>>>> >>>>>>>>> Thank you. >>>>>>>>> Let me try. >>>>>>>>> >>>>>>>>> 2016-06-28 22:18 GMT+09:00 Jonathan Esterhazy < >>>>>>>>> jonathan.esterh...@gmail.com>: >>>>>>>>> >>>>>>>>>> Hyung, >>>>>>>>>> >>>>>>>>>> Yes, here they are. >>>>>>>>>> >>>>>>>>>> zeppelin-env.sh: >>>>>>>>>> >>>>>>>>>> export ZEPPELIN_PORT=8890 >>>>>>>>>> export ZEPPELIN_CONF_DIR=/etc/zeppelin/conf >>>>>>>>>> export ZEPPELIN_LOG_DIR=/var/log/zeppelin >>>>>>>>>> export ZEPPELIN_PID_DIR=/var/run/zeppelin >>>>>>>>>> export ZEPPELIN_PID=$ZEPPELIN_PID_DIR/zeppelin.pid >>>>>>>>>> export ZEPPELIN_NOTEBOOK_DIR=/var/lib/zeppelin/notebook >>>>>>>>>> export ZEPPELIN_WAR_TEMPDIR=/var/run/zeppelin/webapps >>>>>>>>>> export MASTER=yarn-client >>>>>>>>>> export SPARK_HOME=/usr/lib/spark >>>>>>>>>> export HADOOP_CONF_DIR=/etc/hadoop/conf >>>>>>>>>> export >>>>>>>>>> CLASSPATH=":/etc/hive/conf:/usr/lib/hadoop-lzo/lib/*:/usr/lib/hadoop/hadoop-aws.jar:/usr/share/aws/aws-java-sdk/*:/usr/share/aws/emr/emrfs/conf:/usr/share/aws/emr/emrfs/lib/* >>>>>>>>>> :/usr/share/aws/emr/emrfs/auxlib/*" >>>>>>>>>> export JAVA_HOME=/usr/lib/jvm/java-1.8.0 >>>>>>>>>> export ZEPPELIN_NOTEBOOK_S3_BUCKET=mybucket >>>>>>>>>> export ZEPPELIN_NOTEBOOK_S3_USER=zeppelin >>>>>>>>>> export >>>>>>>>>> ZEPPELIN_NOTEBOOK_STORAGE=org.apache.zeppelin.notebook.repo.S3NotebookRepo >>>>>>>>>> >>>>>>>>>> spark-defaults.conf: >>>>>>>>>> >>>>>>>>>> spark.master yarn >>>>>>>>>> spark.driver.extraClassPath >>>>>>>>>> >>>>>>>>>> /etc/hadoop/conf:/etc/hive/conf:/usr/lib/hadoop-lzo/lib/*:/usr/lib/hadoop/hadoop-aws.jar:/usr/share/aws/aws-java-sdk/*:/usr/share/aws/emr/emrfs/conf >>>>>>>>>> :/usr/share/aws/emr/emrfs/lib/*:/usr/share/aws/emr/emrfs/auxlib/* >>>>>>>>>> spark.driver.extraLibraryPath >>>>>>>>>> /usr/lib/hadoop/lib/native:/usr/lib/hadoop-lzo/lib/native >>>>>>>>>> spark.executor.extraClassPath >>>>>>>>>> >>>>>>>>>> /etc/hadoop/conf:/etc/hive/conf:/usr/lib/hadoop-lzo/lib/*:/usr/lib/hadoop/hadoop-aws.jar:/usr/share/aws/aws-java-sdk/*:/usr/share/aws/emr/emrfs/conf >>>>>>>>>> :/usr/share/aws/emr/emrfs/lib/*:/usr/share/aws/emr/emrfs/auxlib/* >>>>>>>>>> spark.executor.extraLibraryPath >>>>>>>>>> /usr/lib/hadoop/lib/native:/usr/lib/hadoop-lzo/lib/native >>>>>>>>>> spark.eventLog.enabled true >>>>>>>>>> spark.eventLog.dir hdfs:///var/log/spark/apps >>>>>>>>>> spark.history.fs.logDirectory hdfs:///var/log/spark/apps >>>>>>>>>> spark.yarn.historyServer.address >>>>>>>>>> ip-172-30-54-30.ec2.internal:18080 >>>>>>>>>> spark.history.ui.port 18080 >>>>>>>>>> spark.shuffle.service.enabled true >>>>>>>>>> spark.driver.extraJavaOptions >>>>>>>>>> -Dlog4j.configuration=file:///etc/spark/conf/log4j.properties >>>>>>>>>> -XX:+UseConcMarkSweepGC -XX:CMSInitiatingOccupancyFraction=70 >>>>>>>>>> -XX:MaxHeapFreeRatio=70 >>>>>>>>>> -XX:+CMSClassUnloadingEnabled -XX:MaxPermSize=512M >>>>>>>>>> -XX:OnOutOfMemoryError='kill -9 %p' >>>>>>>>>> spark.dynamicAllocation.enabled true >>>>>>>>>> spark.executor.extraJavaOptions -verbose:gc -XX:+PrintGCDetails >>>>>>>>>> -XX:+PrintGCDateStamps -XX:+UseConcMarkSweepGC >>>>>>>>>> -XX:CMSInitiatingOccupancyFraction=70 -XX:MaxHeapFreeRatio=70 -XX:+CM >>>>>>>>>> SClassUnloadingEnabled -XX:OnOutOfMemoryError='kill -9 %p' >>>>>>>>>> spark.executor.memory 8640m >>>>>>>>>> spark.executor.cores 7 >>>>>>>>>> spark.authenticate.enableSaslEncryption true >>>>>>>>>> spark.driver.memory 1g >>>>>>>>>> spark.network.sasl.serverAlwaysEncrypt true >>>>>>>>>> spark.driver.cores 1 >>>>>>>>>> spark.ssl.protocol TLSv1.2 >>>>>>>>>> spark.ssl.keyStorePassword password >>>>>>>>>> spark.yarn.maxAppAttempts 1 >>>>>>>>>> spark.ssl.keyStore /etc/emr/security/keystore.jks >>>>>>>>>> spark.authenticate true >>>>>>>>>> spark.ssl.keyPassword password >>>>>>>>>> spark.ssl.enabled true >>>>>>>>>> spark.ssl.enabledAlgorithms TLS_RSA_WITH_AES_256_CBC_SHA >>>>>>>>>> spark.ssl.trustStore /etc/emr/security/truststore.jks >>>>>>>>>> spark.authenticate.secret secret >>>>>>>>>> spark.ssl.trustStorePassword password >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> On Mon, Jun 27, 2016 at 7:33 PM, Hyung Sung Shim < >>>>>>>>>> hss...@nflabs.com> wrote: >>>>>>>>>> >>>>>>>>>>> Hi. >>>>>>>>>>> Could you share your conf/zeppelin-env.sh >>>>>>>>>>> and spark-defaults.conf ? >>>>>>>>>>> >>>>>>>>>>> 2016-06-28 8:52 GMT+09:00 Jonathan Esterhazy < >>>>>>>>>>> jonathan.esterh...@gmail.com>: >>>>>>>>>>> >>>>>>>>>>>> I am having trouble using zeppelin in a spark cluster that has >>>>>>>>>>>> spark node authentication turned on (e.g. with >>>>>>>>>>>> spark.authenticate=true, >>>>>>>>>>>> spark.authenticate.secret=...) >>>>>>>>>>>> >>>>>>>>>>>> Notebook code that calls built-in spark functions (or other >>>>>>>>>>>> things on executor classpath) work fine, but functions defined in >>>>>>>>>>>> the >>>>>>>>>>>> notebook (anonymous or named) throw ClassNotFoundExceptions when >>>>>>>>>>>> called >>>>>>>>>>>> from an executor. >>>>>>>>>>>> >>>>>>>>>>>> For example, this code works: >>>>>>>>>>>> >>>>>>>>>>>> val rdd = sc.textFile("hdfs://my-text-file") >>>>>>>>>>>> rdd.take(1).foreach(println) >>>>>>>>>>>> >>>>>>>>>>>> rdd.saveAsTextFile("hdfs:///my-other-text-file") >>>>>>>>>>>> >>>>>>>>>>>> but code like this... >>>>>>>>>>>> >>>>>>>>>>>> rdd.filter(_.contains("my data")) >>>>>>>>>>>> >>>>>>>>>>>> fails with >>>>>>>>>>>> >>>>>>>>>>>> Caused by: java.lang.ClassNotFoundException: >>>>>>>>>>>> $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$anonfun$1 >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.spark.repl.ExecutorClassLoader.findClass(ExecutorClassLoader.scala:84) >>>>>>>>>>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:424) >>>>>>>>>>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:357) >>>>>>>>>>>> at java.lang.Class.forName0(Native Method) >>>>>>>>>>>> at java.lang.Class.forName(Class.java:348) >>>>>>>>>>>> ... >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> I get the same kind of error if the filter function is defined >>>>>>>>>>>> as a named function in the notebook, or as a member of singleton >>>>>>>>>>>> object >>>>>>>>>>>> defined in the notebook. >>>>>>>>>>>> >>>>>>>>>>>> When I look at the executor's log output, I see this error: >>>>>>>>>>>> >>>>>>>>>>>> 16/06/27 21:36:23 ERROR repl.ExecutorClassLoader: Failed to >>>>>>>>>>>> check existence of class >>>>>>>>>>>> $line31.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$anonfun$1 >>>>>>>>>>>> on REPL >>>>>>>>>>>> class server at https://172.30.54.30:34980 >>>>>>>>>>>> java.lang.NullPointerException >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.spark.repl.ExecutorClassLoader.getClassFileInputStreamFromHttpServer(ExecutorClassLoader.scala:113) >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.spark.repl.ExecutorClassLoader.findClassLocally(ExecutorClassLoader.scala:146) >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.spark.repl.ExecutorClassLoader.findClass(ExecutorClassLoader.scala:76) >>>>>>>>>>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:424) >>>>>>>>>>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:357) >>>>>>>>>>>> at java.lang.Class.forName0(Native Method) >>>>>>>>>>>> at java.lang.Class.forName(Class.java:348) >>>>>>>>>>>> >>>>>>>>>>>> ... >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> If I disable spark authentication, everything works as >>>>>>>>>>>> expected. I am running zeppelin 0.5.6 on spark 1.6.1 with yarn. >>>>>>>>>>>> >>>>>>>>>>>> Has anyone been able to get zeppelin working with spark >>>>>>>>>>>> authentication? >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>> >>>>>> >>>>> >>>> >>> >> >