Hi,

I have tried this example after
https://github.com/apache/incubator-zeppelin/pull/270.

But it is not working for several reasons:

1/ Zeppelin context is not found (!):
val s = z.input("Foo")
<console>:21: error: not found: value z

2/ If I include my jar, the classpath is not communicated to slaves, the
code works only locally (it used to work on the cluster before this
change), I guess there is something wrong with the way I set the classpath
(which is also probably linked to 1/)

I have added this line in zeppelin-env.sh to use one of my jar
export ZEPPELIN_JAVA_OPTS="-Dspark.driver.host=`hostname`
-Dspark.mesos.coarse=true -Dspark.executor.memory=20g -Dspark.cores.max=80
-Dspark.jars=${SOME_JAR} -cp ${SOME_CLASSPATH_FOR_THE_JAR}

How can one add its extraclasspath jar with this new version? Could you add
an ZEPPELIN_EXTRA_JAR, ZEPPELIN_EXTRA_CLASSPATH to zeppelin-env.sh so that
the user can add easily his code?

Best,

David

On Thu, Sep 3, 2015 at 9:39 AM, David Salinas <david.salinas....@gmail.com>
wrote:

> Hi Moon,
>
> Thanks for your reactivity, I will notice you of the result as soon as I
> can.
>
> Best,
>
> David
>
> On Wed, Sep 2, 2015 at 6:51 AM, moon soo Lee <m...@apache.org> wrote:
>
>> Hi,
>>
>> I just pushed patch for ZEPPELIN-262 at
>> https://github.com/apache/incubator-zeppelin/pull/270.
>> It'll take some time to be reviewed and merged into master.
>> Before that, you can try the branch of the PullRequest.
>>
>> I believe it'll solve your problem, but let me know when you still have
>> problem after this patch.
>>
>> Thanks,
>> moon
>>
>> On Tue, Sep 1, 2015 at 2:46 PM moon soo Lee <m...@apache.org> wrote:
>>
>>> Hi,
>>>
>>> I'm testing patch for ZEPPELIN-262 with some environments that i have.
>>> I think i can create pullrequest tonight.
>>>
>>> Thanks,
>>> moon
>>>
>>>
>>> On Tue, Sep 1, 2015 at 1:34 PM Steven Kirtzic <
>>> steven.kirtzic.f...@statefarm.com> wrote:
>>>
>>>> Hi Moon,
>>>>
>>>>
>>>>
>>>> When are you guys targeting the release for Zeppelin-262? Thanks,
>>>>
>>>>
>>>>
>>>> -Steven
>>>>
>>>>
>>>>
>>>> *From:* moon soo Lee [mailto:m...@apache.org]
>>>> *Sent:* Tuesday, September 01, 2015 12:38 AM
>>>> *To:* users@zeppelin.incubator.apache.org
>>>> *Subject:* Re: Closure issue with spark 1.4.1
>>>>
>>>>
>>>>
>>>> Hi David, Jerry,
>>>>
>>>>
>>>>
>>>> There're series of efforts to improve spark integration.
>>>>
>>>>
>>>>
>>>> Work with provided version of Spark
>>>>
>>>> https://issues.apache.org/jira/browse/ZEPPELIN-160
>>>>
>>>>
>>>>
>>>> Self diagnostics of configuration
>>>>
>>>> https://issues.apache.org/jira/browse/ZEPPELIN-256
>>>>
>>>>
>>>>
>>>> Use spark-submit to run spark interpreter process
>>>>
>>>> https://issues.apache.org/jira/browse/ZEPPELIN-262
>>>>
>>>>
>>>>
>>>> I saw many people struggled with configuring spark in Zeppelin with
>>>> various environments in the mailing list.
>>>>
>>>> ZEPPELIN-262 will virtually solve all the problems around configuration
>>>> with Spark.
>>>>
>>>>
>>>>
>>>> Thanks for sharing your problems and feedback. That enables zeppelin
>>>> make progress.
>>>>
>>>>
>>>>
>>>> Best,
>>>>
>>>> moon
>>>>
>>>>
>>>>
>>>> On Mon, Aug 31, 2015 at 9:17 PM Jerry Lam <chiling...@gmail.com> wrote:
>>>>
>>>> Hi David,
>>>>
>>>>
>>>>
>>>> We gave up on zeppelin because of the lack of support. It seems that
>>>> zeppelin has a lot of fancy features but lack of depth. Only time will tell
>>>> if zeppelin can overcome those limitations.
>>>>
>>>>
>>>>
>>>> Good luck,
>>>>
>>>>
>>>>
>>>> Jerry
>>>>
>>>>
>>>>
>>>> On Mon, Aug 31, 2015 at 8:17 AM, David Salinas <
>>>> david.salinas....@gmail.com> wrote:
>>>>
>>>> Hi all,
>>>>
>>>> Has anyone been able to reproduce the error with the last code snipplet
>>>> I gave? It fails 100% of the time on cluster for me.
>>>> This serialization issue asking for ZeppelinContext comes also in many
>>>> other cases in my setting where it should not be the case as it works fine
>>>> with spark shell.
>>>>
>>>> Best regards,
>>>>
>>>> David
>>>>
>>>>
>>>>
>>>> On Mon, Aug 24, 2015 at 9:07 PM, Jerry Lam <chiling...@gmail.com>
>>>> wrote:
>>>>
>>>> Hi Zeppelin developers,
>>>>
>>>>
>>>>
>>>> This issue sounds very serious. Is this specific to David's use case
>>>> here?
>>>>
>>>>
>>>>
>>>> Best Regards,
>>>>
>>>>
>>>>
>>>> Jerry
>>>>
>>>>
>>>>
>>>> On Mon, Aug 24, 2015 at 1:28 PM, David Salinas <
>>>> david.salinas....@gmail.com> wrote:
>>>>
>>>> I have looked at the SparkInterpreter.java code and this is indeed the
>>>> issue. Whenever one uses an instruction with z.input("...") something then
>>>> no spark transformation can work as z will be shipped to the slaves where
>>>> Zeppelin is not installed as showed by the example I sent.
>>>>
>>>> A workaround could be to interpret separately the variables (by
>>>> defining a map of variables before interpreting).
>>>>
>>>> Best,
>>>>
>>>> David
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> On Mon, Aug 24, 2015 at 6:45 PM, David Salinas <
>>>> david.salinas....@gmail.com> wrote:
>>>>
>>>> Hi Moon,
>>>>
>>>> I found another way to reproduce the problem:
>>>>
>>>> //cell 1 does not work
>>>>
>>>> val file = "hdfs://someclusterfile.json"
>>>> val s = z.input("Foo").toString
>>>> val textFile = sc.textFile(file)
>>>> textFile.filter(_.contains(s)).count
>>>> //org.apache.spark.SparkException: Job aborted due to stage failure:
>>>> Task 41 in stage 5.0 failed 4 times, most recent failure: Lost task 41.3 in
>>>> stage 5.0 (TID 2735,XXX.com ): java.lang.NoClassDefFoundError:
>>>> Lorg/apache/zeppelin/spark/ZeppelinContext;
>>>>
>>>> // cell 2 works
>>>>
>>>> val file = "hdfs://someclusterfile.json"
>>>> val s = "Y"
>>>> val textFile = sc.textFile(file)
>>>> textFile.filter(_.contains(s)).count
>>>> //res19: Long = 109
>>>>
>>>> This kind of issue happens often also when using variables from other
>>>> cells and also when taking closure for transformation. Maybe you are
>>>> reading variables inside the transformation with something like
>>>> "z.get("s")" which causes z to be send to the slaves as one of its member
>>>> is used (although I also sometimes have this issue without using anything
>>>> from other cells).
>>>>
>>>> Best,
>>>>
>>>> David
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> On Mon, Aug 24, 2015 at 10:34 AM, David Salinas <
>>>> david.salinas....@gmail.com> wrote:
>>>>
>>>> Sorry I forgot to mention my environment:
>>>>
>>>> mesos 0.17, spark 1.4.1, scala 2.10.4, java 1.8
>>>>
>>>>
>>>>
>>>> On Mon, Aug 24, 2015 at 10:32 AM, David Salinas <
>>>> david.salinas....@gmail.com> wrote:
>>>>
>>>> Hi Moon,
>>>>
>>>>
>>>> Today I cannot reproduce the bug with elementary example either but it
>>>> is still impacting all my notebooks. The weird thing is that when calling a
>>>> transformation with map, it takes Zeppelin Context in the closure which
>>>> gives these java.lang.NoClassDefFoundError:
>>>> Lorg/apache/zeppelin/spark/ZeppelinContext errors (spark shell run this
>>>> command without any problem). I will try to find another example that is
>>>> more persistent (it is weird this example was failing yesterday). Do you
>>>> have any idea of what could cause Zeppelin Context to be included in the
>>>> closure?
>>>>
>>>> Best,
>>>>
>>>> David
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> On Fri, Aug 21, 2015 at 6:29 PM, moon soo Lee <m...@apache.org> wrote:
>>>>
>>>> I have tested your code and can not reproduce the problem.
>>>>
>>>>
>>>>
>>>> Could you share your environment? how did you configure Zeppelin with
>>>> Spark?
>>>>
>>>>
>>>>
>>>> Thanks,
>>>>
>>>> moon
>>>>
>>>>
>>>>
>>>> On Fri, Aug 21, 2015 at 2:25 AM David Salinas <
>>>> david.salinas....@gmail.com> wrote:
>>>>
>>>> Hi,
>>>>
>>>> I have a problem when using spark closure. This error was not appearing
>>>> with spark 1.2.1.
>>>>
>>>> I have included a reproducible example that happens when taking the
>>>> closure (Zeppelin has been built with head of master with this command mvn
>>>> install -DskipTests -Pspark-1.4 -Dspark.version=1.4.1
>>>> -Dhadoop.version=2.2.0 -Dprotobuf.version=2.5.0). Does anyone ever
>>>> encountered this problem? All my previous notebooks are broken by this :(
>>>>
>>>> ------------------------------
>>>> val textFile = sc.textFile("hdfs://somefile.txt")
>>>>
>>>> val f = (s: String) => s+s
>>>> textFile.map(f).count
>>>> //works fine
>>>> //res145: Long = 407
>>>>
>>>>
>>>> def f(s:String) = {
>>>>     s+s
>>>> }
>>>> textFile.map(f).count
>>>>
>>>> //fails ->
>>>>
>>>> org.apache.spark.SparkException: Job aborted due to stage failure: Task
>>>> 566 in stage 87.0 failed 4 times, most recent failure: Lost task 566.3 in
>>>> stage 87.0 (TID 43396, XXX.com): java.lang.NoClassDefFoundError:
>>>> Lorg/apache/zeppelin/spark/ZeppelinContext; at
>>>> java.lang.Class.getDeclaredFields0(Native Method) at
>>>> java.lang.Class.privateGetDeclaredFields(Class.java:2583) at
>>>> java.lang.Class.getDeclaredField(Class.java:2068) ...
>>>> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1924)
>>>> at
>>>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1801)
>>>> at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1351) at
>>>> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:2000) at
>>>>
>>>> Best,
>>>>
>>>> David
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>

Reply via email to