Hi Moon, Thanks for your reactivity, I will notice you of the result as soon as I can.
Best, David On Wed, Sep 2, 2015 at 6:51 AM, moon soo Lee <m...@apache.org> wrote: > Hi, > > I just pushed patch for ZEPPELIN-262 at > https://github.com/apache/incubator-zeppelin/pull/270. > It'll take some time to be reviewed and merged into master. > Before that, you can try the branch of the PullRequest. > > I believe it'll solve your problem, but let me know when you still have > problem after this patch. > > Thanks, > moon > > On Tue, Sep 1, 2015 at 2:46 PM moon soo Lee <m...@apache.org> wrote: > >> Hi, >> >> I'm testing patch for ZEPPELIN-262 with some environments that i have. >> I think i can create pullrequest tonight. >> >> Thanks, >> moon >> >> >> On Tue, Sep 1, 2015 at 1:34 PM Steven Kirtzic < >> steven.kirtzic.f...@statefarm.com> wrote: >> >>> Hi Moon, >>> >>> >>> >>> When are you guys targeting the release for Zeppelin-262? Thanks, >>> >>> >>> >>> -Steven >>> >>> >>> >>> *From:* moon soo Lee [mailto:m...@apache.org] >>> *Sent:* Tuesday, September 01, 2015 12:38 AM >>> *To:* users@zeppelin.incubator.apache.org >>> *Subject:* Re: Closure issue with spark 1.4.1 >>> >>> >>> >>> Hi David, Jerry, >>> >>> >>> >>> There're series of efforts to improve spark integration. >>> >>> >>> >>> Work with provided version of Spark >>> >>> https://issues.apache.org/jira/browse/ZEPPELIN-160 >>> >>> >>> >>> Self diagnostics of configuration >>> >>> https://issues.apache.org/jira/browse/ZEPPELIN-256 >>> >>> >>> >>> Use spark-submit to run spark interpreter process >>> >>> https://issues.apache.org/jira/browse/ZEPPELIN-262 >>> >>> >>> >>> I saw many people struggled with configuring spark in Zeppelin with >>> various environments in the mailing list. >>> >>> ZEPPELIN-262 will virtually solve all the problems around configuration >>> with Spark. >>> >>> >>> >>> Thanks for sharing your problems and feedback. That enables zeppelin >>> make progress. >>> >>> >>> >>> Best, >>> >>> moon >>> >>> >>> >>> On Mon, Aug 31, 2015 at 9:17 PM Jerry Lam <chiling...@gmail.com> wrote: >>> >>> Hi David, >>> >>> >>> >>> We gave up on zeppelin because of the lack of support. It seems that >>> zeppelin has a lot of fancy features but lack of depth. Only time will tell >>> if zeppelin can overcome those limitations. >>> >>> >>> >>> Good luck, >>> >>> >>> >>> Jerry >>> >>> >>> >>> On Mon, Aug 31, 2015 at 8:17 AM, David Salinas < >>> david.salinas....@gmail.com> wrote: >>> >>> Hi all, >>> >>> Has anyone been able to reproduce the error with the last code snipplet >>> I gave? It fails 100% of the time on cluster for me. >>> This serialization issue asking for ZeppelinContext comes also in many >>> other cases in my setting where it should not be the case as it works fine >>> with spark shell. >>> >>> Best regards, >>> >>> David >>> >>> >>> >>> On Mon, Aug 24, 2015 at 9:07 PM, Jerry Lam <chiling...@gmail.com> wrote: >>> >>> Hi Zeppelin developers, >>> >>> >>> >>> This issue sounds very serious. Is this specific to David's use case >>> here? >>> >>> >>> >>> Best Regards, >>> >>> >>> >>> Jerry >>> >>> >>> >>> On Mon, Aug 24, 2015 at 1:28 PM, David Salinas < >>> david.salinas....@gmail.com> wrote: >>> >>> I have looked at the SparkInterpreter.java code and this is indeed the >>> issue. Whenever one uses an instruction with z.input("...") something then >>> no spark transformation can work as z will be shipped to the slaves where >>> Zeppelin is not installed as showed by the example I sent. >>> >>> A workaround could be to interpret separately the variables (by defining >>> a map of variables before interpreting). >>> >>> Best, >>> >>> David >>> >>> >>> >>> >>> >>> On Mon, Aug 24, 2015 at 6:45 PM, David Salinas < >>> david.salinas....@gmail.com> wrote: >>> >>> Hi Moon, >>> >>> I found another way to reproduce the problem: >>> >>> //cell 1 does not work >>> >>> val file = "hdfs://someclusterfile.json" >>> val s = z.input("Foo").toString >>> val textFile = sc.textFile(file) >>> textFile.filter(_.contains(s)).count >>> //org.apache.spark.SparkException: Job aborted due to stage failure: >>> Task 41 in stage 5.0 failed 4 times, most recent failure: Lost task 41.3 in >>> stage 5.0 (TID 2735,XXX.com ): java.lang.NoClassDefFoundError: >>> Lorg/apache/zeppelin/spark/ZeppelinContext; >>> >>> // cell 2 works >>> >>> val file = "hdfs://someclusterfile.json" >>> val s = "Y" >>> val textFile = sc.textFile(file) >>> textFile.filter(_.contains(s)).count >>> //res19: Long = 109 >>> >>> This kind of issue happens often also when using variables from other >>> cells and also when taking closure for transformation. Maybe you are >>> reading variables inside the transformation with something like >>> "z.get("s")" which causes z to be send to the slaves as one of its member >>> is used (although I also sometimes have this issue without using anything >>> from other cells). >>> >>> Best, >>> >>> David >>> >>> >>> >>> >>> >>> On Mon, Aug 24, 2015 at 10:34 AM, David Salinas < >>> david.salinas....@gmail.com> wrote: >>> >>> Sorry I forgot to mention my environment: >>> >>> mesos 0.17, spark 1.4.1, scala 2.10.4, java 1.8 >>> >>> >>> >>> On Mon, Aug 24, 2015 at 10:32 AM, David Salinas < >>> david.salinas....@gmail.com> wrote: >>> >>> Hi Moon, >>> >>> >>> Today I cannot reproduce the bug with elementary example either but it >>> is still impacting all my notebooks. The weird thing is that when calling a >>> transformation with map, it takes Zeppelin Context in the closure which >>> gives these java.lang.NoClassDefFoundError: >>> Lorg/apache/zeppelin/spark/ZeppelinContext errors (spark shell run this >>> command without any problem). I will try to find another example that is >>> more persistent (it is weird this example was failing yesterday). Do you >>> have any idea of what could cause Zeppelin Context to be included in the >>> closure? >>> >>> Best, >>> >>> David >>> >>> >>> >>> >>> >>> On Fri, Aug 21, 2015 at 6:29 PM, moon soo Lee <m...@apache.org> wrote: >>> >>> I have tested your code and can not reproduce the problem. >>> >>> >>> >>> Could you share your environment? how did you configure Zeppelin with >>> Spark? >>> >>> >>> >>> Thanks, >>> >>> moon >>> >>> >>> >>> On Fri, Aug 21, 2015 at 2:25 AM David Salinas < >>> david.salinas....@gmail.com> wrote: >>> >>> Hi, >>> >>> I have a problem when using spark closure. This error was not appearing >>> with spark 1.2.1. >>> >>> I have included a reproducible example that happens when taking the >>> closure (Zeppelin has been built with head of master with this command mvn >>> install -DskipTests -Pspark-1.4 -Dspark.version=1.4.1 >>> -Dhadoop.version=2.2.0 -Dprotobuf.version=2.5.0). Does anyone ever >>> encountered this problem? All my previous notebooks are broken by this :( >>> >>> ------------------------------ >>> val textFile = sc.textFile("hdfs://somefile.txt") >>> >>> val f = (s: String) => s+s >>> textFile.map(f).count >>> //works fine >>> //res145: Long = 407 >>> >>> >>> def f(s:String) = { >>> s+s >>> } >>> textFile.map(f).count >>> >>> //fails -> >>> >>> org.apache.spark.SparkException: Job aborted due to stage failure: Task >>> 566 in stage 87.0 failed 4 times, most recent failure: Lost task 566.3 in >>> stage 87.0 (TID 43396, XXX.com): java.lang.NoClassDefFoundError: >>> Lorg/apache/zeppelin/spark/ZeppelinContext; at >>> java.lang.Class.getDeclaredFields0(Native Method) at >>> java.lang.Class.privateGetDeclaredFields(Class.java:2583) at >>> java.lang.Class.getDeclaredField(Class.java:2068) ... >>> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1924) at >>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1801) >>> at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1351) at >>> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:2000) at >>> >>> Best, >>> >>> David >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>>