I just tried this with using zeppelin master, and the classnotfound error
is still there.


On Thu, Jul 7, 2016 at 7:07 AM, Hyung Sung Shim <hss...@nflabs.com> wrote:

> Hi. Jonathan.
> If it is possible, could you use the zeppelin master on the github?
> Zeppelin is working well with spark authentication when i tested.
>
>
>
> 2016-07-02 3:05 GMT+09:00 Jonathan Esterhazy <jonathan.esterh...@gmail.com
> >:
>
>> Hyung, thx for your help. I opened these:
>>
>> https://issues.apache.org/jira/browse/ZEPPELIN-1096 (this scala problem)
>> https://issues.apache.org/jira/browse/ZEPPELIN-1097 (similar looking
>> python problem)
>>
>> LMK if I can provide more info or help in some way.
>>
>> On Fri, Jul 1, 2016 at 5:08 AM, Hyung Sung Shim <hss...@nflabs.com>
>> wrote:
>>
>>> Hi Jonathan.
>>> Unfortunately I got same error in my test bed.
>>> Do you mind create an jira issue for this?
>>>
>>> 2016-07-01 3:19 GMT+09:00 Jonathan Esterhazy <
>>> jonathan.esterh...@gmail.com>:
>>>
>>>> I added this and still get the same exception. The same property is
>>>> also set in spark-defaults.conf.
>>>>
>>>> After that didn't work, I also tried adding --conf
>>>> spark.authenticate=true and --conf
>>>> spark.authenticate.enableSaslEncryption=true, to match the other related
>>>> settings in spark-defaults.conf. Still get the same classnotfoundexception.
>>>>
>>>>
>>>> On Thu, Jun 30, 2016 at 10:45 AM, Hyung Sung Shim <hss...@nflabs.com>
>>>> wrote:
>>>>
>>>>> Please add the *export SPARK_SUBMIT_OPTIONS="--conf
>>>>> spark.authenticate.secret=secret"* in conf/zeppelin-env.sh and
>>>>> restart zeppelin and retry your code.
>>>>>
>>>>>
>>>>> 2016-06-30 23:34 GMT+09:00 Jonathan Esterhazy <
>>>>> jonathan.esterh...@gmail.com>:
>>>>>
>>>>>> yes it does. i only see this problem in zeppelin.
>>>>>>
>>>>>> On Thu, Jun 30, 2016 at 7:05 AM, Hyung Sung Shim <hss...@nflabs.com>
>>>>>> wrote:
>>>>>>
>>>>>>> Hi Jonathan.
>>>>>>> It's not easy to build the test environments but I am working on
>>>>>>> this.
>>>>>>> I have question for you.
>>>>>>> Does your code working well on spark-shell in the spark.authenticate
>>>>>>> mode?
>>>>>>>
>>>>>>> 2016-06-30 22:47 GMT+09:00 Jonathan Esterhazy <
>>>>>>> jonathan.esterh...@gmail.com>:
>>>>>>>
>>>>>>>> Hyung, did you have any luck w/ zeppelin + spark authentication?
>>>>>>>> I'm quite stumped.
>>>>>>>>
>>>>>>>> thx.
>>>>>>>>
>>>>>>>> On Tue, Jun 28, 2016 at 9:11 PM, Hyung Sung Shim <hss...@nflabs.com
>>>>>>>> > wrote:
>>>>>>>>
>>>>>>>>> Thank you.
>>>>>>>>> Let me try.
>>>>>>>>>
>>>>>>>>> 2016-06-28 22:18 GMT+09:00 Jonathan Esterhazy <
>>>>>>>>> jonathan.esterh...@gmail.com>:
>>>>>>>>>
>>>>>>>>>> Hyung,
>>>>>>>>>>
>>>>>>>>>> Yes, here they are.
>>>>>>>>>>
>>>>>>>>>> zeppelin-env.sh:
>>>>>>>>>>
>>>>>>>>>> export ZEPPELIN_PORT=8890
>>>>>>>>>> export ZEPPELIN_CONF_DIR=/etc/zeppelin/conf
>>>>>>>>>> export ZEPPELIN_LOG_DIR=/var/log/zeppelin
>>>>>>>>>> export ZEPPELIN_PID_DIR=/var/run/zeppelin
>>>>>>>>>> export ZEPPELIN_PID=$ZEPPELIN_PID_DIR/zeppelin.pid
>>>>>>>>>> export ZEPPELIN_NOTEBOOK_DIR=/var/lib/zeppelin/notebook
>>>>>>>>>> export ZEPPELIN_WAR_TEMPDIR=/var/run/zeppelin/webapps
>>>>>>>>>> export MASTER=yarn-client
>>>>>>>>>> export SPARK_HOME=/usr/lib/spark
>>>>>>>>>> export HADOOP_CONF_DIR=/etc/hadoop/conf
>>>>>>>>>> export
>>>>>>>>>> CLASSPATH=":/etc/hive/conf:/usr/lib/hadoop-lzo/lib/*:/usr/lib/hadoop/hadoop-aws.jar:/usr/share/aws/aws-java-sdk/*:/usr/share/aws/emr/emrfs/conf:/usr/share/aws/emr/emrfs/lib/*
>>>>>>>>>> :/usr/share/aws/emr/emrfs/auxlib/*"
>>>>>>>>>> export JAVA_HOME=/usr/lib/jvm/java-1.8.0
>>>>>>>>>> export ZEPPELIN_NOTEBOOK_S3_BUCKET=mybucket
>>>>>>>>>> export ZEPPELIN_NOTEBOOK_S3_USER=zeppelin
>>>>>>>>>> export
>>>>>>>>>> ZEPPELIN_NOTEBOOK_STORAGE=org.apache.zeppelin.notebook.repo.S3NotebookRepo
>>>>>>>>>>
>>>>>>>>>> spark-defaults.conf:
>>>>>>>>>>
>>>>>>>>>> spark.master                     yarn
>>>>>>>>>> spark.driver.extraClassPath
>>>>>>>>>>  
>>>>>>>>>> /etc/hadoop/conf:/etc/hive/conf:/usr/lib/hadoop-lzo/lib/*:/usr/lib/hadoop/hadoop-aws.jar:/usr/share/aws/aws-java-sdk/*:/usr/share/aws/emr/emrfs/conf
>>>>>>>>>> :/usr/share/aws/emr/emrfs/lib/*:/usr/share/aws/emr/emrfs/auxlib/*
>>>>>>>>>> spark.driver.extraLibraryPath
>>>>>>>>>>  /usr/lib/hadoop/lib/native:/usr/lib/hadoop-lzo/lib/native
>>>>>>>>>> spark.executor.extraClassPath
>>>>>>>>>>  
>>>>>>>>>> /etc/hadoop/conf:/etc/hive/conf:/usr/lib/hadoop-lzo/lib/*:/usr/lib/hadoop/hadoop-aws.jar:/usr/share/aws/aws-java-sdk/*:/usr/share/aws/emr/emrfs/conf
>>>>>>>>>> :/usr/share/aws/emr/emrfs/lib/*:/usr/share/aws/emr/emrfs/auxlib/*
>>>>>>>>>> spark.executor.extraLibraryPath
>>>>>>>>>>  /usr/lib/hadoop/lib/native:/usr/lib/hadoop-lzo/lib/native
>>>>>>>>>> spark.eventLog.enabled           true
>>>>>>>>>> spark.eventLog.dir               hdfs:///var/log/spark/apps
>>>>>>>>>> spark.history.fs.logDirectory    hdfs:///var/log/spark/apps
>>>>>>>>>> spark.yarn.historyServer.address
>>>>>>>>>> ip-172-30-54-30.ec2.internal:18080
>>>>>>>>>> spark.history.ui.port            18080
>>>>>>>>>> spark.shuffle.service.enabled    true
>>>>>>>>>> spark.driver.extraJavaOptions
>>>>>>>>>>  -Dlog4j.configuration=file:///etc/spark/conf/log4j.properties
>>>>>>>>>> -XX:+UseConcMarkSweepGC -XX:CMSInitiatingOccupancyFraction=70
>>>>>>>>>> -XX:MaxHeapFreeRatio=70
>>>>>>>>>> -XX:+CMSClassUnloadingEnabled -XX:MaxPermSize=512M
>>>>>>>>>> -XX:OnOutOfMemoryError='kill -9 %p'
>>>>>>>>>> spark.dynamicAllocation.enabled  true
>>>>>>>>>> spark.executor.extraJavaOptions  -verbose:gc -XX:+PrintGCDetails
>>>>>>>>>> -XX:+PrintGCDateStamps -XX:+UseConcMarkSweepGC
>>>>>>>>>> -XX:CMSInitiatingOccupancyFraction=70 -XX:MaxHeapFreeRatio=70 -XX:+CM
>>>>>>>>>> SClassUnloadingEnabled -XX:OnOutOfMemoryError='kill -9 %p'
>>>>>>>>>> spark.executor.memory            8640m
>>>>>>>>>> spark.executor.cores             7
>>>>>>>>>> spark.authenticate.enableSaslEncryption true
>>>>>>>>>> spark.driver.memory              1g
>>>>>>>>>> spark.network.sasl.serverAlwaysEncrypt true
>>>>>>>>>> spark.driver.cores               1
>>>>>>>>>> spark.ssl.protocol               TLSv1.2
>>>>>>>>>> spark.ssl.keyStorePassword       password
>>>>>>>>>> spark.yarn.maxAppAttempts        1
>>>>>>>>>> spark.ssl.keyStore               /etc/emr/security/keystore.jks
>>>>>>>>>> spark.authenticate               true
>>>>>>>>>> spark.ssl.keyPassword            password
>>>>>>>>>> spark.ssl.enabled                true
>>>>>>>>>> spark.ssl.enabledAlgorithms      TLS_RSA_WITH_AES_256_CBC_SHA
>>>>>>>>>> spark.ssl.trustStore             /etc/emr/security/truststore.jks
>>>>>>>>>> spark.authenticate.secret        secret
>>>>>>>>>> spark.ssl.trustStorePassword     password
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> On Mon, Jun 27, 2016 at 7:33 PM, Hyung Sung Shim <
>>>>>>>>>> hss...@nflabs.com> wrote:
>>>>>>>>>>
>>>>>>>>>>> Hi.
>>>>>>>>>>> Could you share your conf/zeppelin-env.sh
>>>>>>>>>>> and spark-defaults.conf ?
>>>>>>>>>>>
>>>>>>>>>>> 2016-06-28 8:52 GMT+09:00 Jonathan Esterhazy <
>>>>>>>>>>> jonathan.esterh...@gmail.com>:
>>>>>>>>>>>
>>>>>>>>>>>> I am having trouble using zeppelin in a spark cluster that has
>>>>>>>>>>>> spark node authentication turned on (e.g. with 
>>>>>>>>>>>> spark.authenticate=true,
>>>>>>>>>>>> spark.authenticate.secret=...)
>>>>>>>>>>>>
>>>>>>>>>>>> Notebook code that calls built-in spark functions (or other
>>>>>>>>>>>> things on executor classpath) work fine, but functions defined in 
>>>>>>>>>>>> the
>>>>>>>>>>>> notebook (anonymous or named) throw ClassNotFoundExceptions when 
>>>>>>>>>>>> called
>>>>>>>>>>>> from an executor.
>>>>>>>>>>>>
>>>>>>>>>>>> For example, this code works:
>>>>>>>>>>>>
>>>>>>>>>>>> val rdd = sc.textFile("hdfs://my-text-file")
>>>>>>>>>>>> rdd.take(1).foreach(println)
>>>>>>>>>>>>
>>>>>>>>>>>> rdd.saveAsTextFile("hdfs:///my-other-text-file")
>>>>>>>>>>>>
>>>>>>>>>>>> but code like this...
>>>>>>>>>>>>
>>>>>>>>>>>> rdd.filter(_.contains("my data"))
>>>>>>>>>>>>
>>>>>>>>>>>> fails with
>>>>>>>>>>>>
>>>>>>>>>>>> Caused by: java.lang.ClassNotFoundException:
>>>>>>>>>>>> $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$anonfun$1
>>>>>>>>>>>> at
>>>>>>>>>>>> org.apache.spark.repl.ExecutorClassLoader.findClass(ExecutorClassLoader.scala:84)
>>>>>>>>>>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:424)
>>>>>>>>>>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:357)
>>>>>>>>>>>> at java.lang.Class.forName0(Native Method)
>>>>>>>>>>>> at java.lang.Class.forName(Class.java:348)
>>>>>>>>>>>>         ...
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> I get the same kind of error if the filter function is defined
>>>>>>>>>>>> as a named function in the notebook, or as a member of singleton 
>>>>>>>>>>>> object
>>>>>>>>>>>> defined in the notebook.
>>>>>>>>>>>>
>>>>>>>>>>>> When I look at the executor's log output, I see this error:
>>>>>>>>>>>>
>>>>>>>>>>>> 16/06/27 21:36:23 ERROR repl.ExecutorClassLoader: Failed to
>>>>>>>>>>>> check existence of class
>>>>>>>>>>>> $line31.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$anonfun$1 
>>>>>>>>>>>> on REPL
>>>>>>>>>>>> class server at https://172.30.54.30:34980
>>>>>>>>>>>> java.lang.NullPointerException
>>>>>>>>>>>> at
>>>>>>>>>>>> org.apache.spark.repl.ExecutorClassLoader.getClassFileInputStreamFromHttpServer(ExecutorClassLoader.scala:113)
>>>>>>>>>>>> at
>>>>>>>>>>>> org.apache.spark.repl.ExecutorClassLoader.findClassLocally(ExecutorClassLoader.scala:146)
>>>>>>>>>>>> at
>>>>>>>>>>>> org.apache.spark.repl.ExecutorClassLoader.findClass(ExecutorClassLoader.scala:76)
>>>>>>>>>>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:424)
>>>>>>>>>>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:357)
>>>>>>>>>>>> at java.lang.Class.forName0(Native Method)
>>>>>>>>>>>> at java.lang.Class.forName(Class.java:348)
>>>>>>>>>>>>
>>>>>>>>>>>>         ...
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> If I disable spark authentication, everything works as
>>>>>>>>>>>> expected. I am running zeppelin 0.5.6 on spark 1.6.1 with yarn.
>>>>>>>>>>>>
>>>>>>>>>>>> Has anyone been able to get zeppelin working with spark
>>>>>>>>>>>> authentication?
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>

Reply via email to