Can you check `export` to see what do you have SPARK_HOME set to?

 
> Date: Sun, 11 Oct 2015 16:50:23 +0530
> Subject: Re: Problem loading data from HDFS
> From: ranveer.k.ku...@gmail.com
> To: us...@zeppelin.incubator.apache.org
> CC: dev@zeppelin.incubator.apache.org
> 
> Still same problem my environment is now exact :
> 
> hadoop : 2.3
> zeppelin : zeppelin-0.5.0-incubating-bin-spark-1.3.1_hadoop-2.3
> spark : spark-1.3.1-bin-hadoop2.3
> 
> while checking version in zeppelin using sc.version, it sowing 1.4.0
> 
> pls help I am stuck..
> 
> thanks
> 
> On Sun, Oct 11, 2015 at 10:58 AM, Ranveer kumar <ranveer.k.ku...@gmail.com>
> wrote:
> 
> > I am using 0.5.0 now but earlier I was using own build with hadoop 2.6
> > with yarn.
> >
> > Is 0.5.0 binary not compatible with hadoop 2.6?
> >
> > I am again going to build zeppelin with hadoop 2.6 also any pre build
> > zeppelin binary for hadoop 2.6 so directly download.
> >
> > Regards
> > On 11 Oct 2015 10:07, "Felix Cheung" <felixcheun...@hotmail.com> wrote:
> >
> >> Are you getting the last Zeppelin release binary from
> >> https://zeppelin.incubator.apache.org/download.html
> >> <https://zeppelin.incubator.apache.org/download.html?> or the Apache
> >> release page?
> >>
> >> It looks like the 0.5.0 release of Zeppelin is only built with Hadoop
> >> 2.3. We might be cutting another release soon or you might need to build
> >> you own...
> >>
> >> _____________________________
> >> From: Ranveer Kumar <ranveer.s...@gmail.com>
> >> Sent: Saturday, October 10, 2015 9:00 PM
> >> Subject: Re: Problem loading data from HDFS
> >> To: <dev@zeppelin.incubator.apache.org>
> >> Cc: <us...@zeppelin.incubator.apache.org>
> >>
> >>
> >> Hi Felix thanks for reply.
> >>
> >> I am using binary download from apache.
> >>
> >> I also tried using 2.6 hadoop, which build from source code.
> >>
> >> Which version of hadoop is compatible with zeppelin binary available on
> >> apache site.
> >> On 11 Oct 2015 09:05, "Felix Cheung" < felixcheun...@hotmail.com> wrote:
> >>
> >>> Is your Zeppelin built with Hadoop 2.6?
> >>>
> >>>
> >>>
> >>>
> >>>
> >>>
> >>> On Sat, Oct 10, 2015 at 7:35 PM -0700, "Ranveer kumar" <
> >>> ranveer.k.ku...@gmail.com> wrote:
> >>> Hi All,
> >>>
> >>> I am new in Zepplin and HDFS. I manage to install zeppelin and working
> >>> fine
> >>> while loading data from local directory . But when same I am trying to
> >>> load
> >>> from HDFS (install locally standalone mode).
> >>>
> >>> here is my code :
> >>>
> >>> val bankText = sc.textFile("/home/ranveer/Desktop/CSVs/bank-full.csv")
> >>>
> >>> above is working fine.
> >>>
> >>> but when trying form hdfs :
> >>>
> >>> val bankText = sc.textFile("hdfs:// 127.0.0.1:9000/demo/
> >>> < http://127.0.0.1:9000/demo/csv/bank-full.csv>csv
> >>> < http://127.0.0.1:9000/demo/csv/bank-full.csv>/
> >>> < http://127.0.0.1:9000/demo/csv/bank-full.csv>bank-full.csv
> >>> < http://127.0.0.1:9000/demo/csv/bank-full.csv>")
> >>>
> >>> not working and giving error :
> >>>
> >>> java.lang.VerifyError: class
> >>> org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$GetAdditionalDatanodeRequestProto
> >>>
> >>> overrides final method
> >>> getUnknownFields.()Lcom/google/protobuf/UnknownFieldSet; at
> >>> java.lang.ClassLoader.defineClass1(Native Method) at
> >>> java.lang.ClassLoader.defineClass(ClassLoader.java:800) at
> >>> java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142)
> >>> at
> >>> java.net.URLClassLoader.defineClass(URLClassLoader.java:449) at
> >>> java.net.URLClassLoader.access$100(URLClassLoader.java:71)
> >>>
> >>> complete code is :
> >>>
> >>> val bankText = sc.textFile("hdfs:// 127.0.0.1:9000/demo/
> >>> < http://127.0.0.1:9000/demo/csv/bank-full.csv>csv
> >>> < http://127.0.0.1:9000/demo/csv/bank-full.csv>/
> >>> < http://127.0.0.1:9000/demo/csv/bank-full.csv>bank-full.csv
> >>> < http://127.0.0.1:9000/demo/csv/bank-full.csv>")
> >>> // val bankText =
> >>> sc.textFile("/home/ranveer/Desktop/CSVs/bank-full.csv")
> >>> case class Bank(age:Integer, job:String, marital : String, education :
> >>> String, balance : Integer)
> >>>
> >>> val bank = bankText.map(s=>s.split(";")).filter(s=>s(0)!="\"age\"").map(
> >>>     s=>Bank(s(0).toInt,
> >>>             s(1).replaceAll("\"", ""),
> >>>             s(2).replaceAll("\"", ""),
> >>>             s(3).replaceAll("\"", ""),
> >>>             s(5).replaceAll("\"", "").toInt
> >>>         )
> >>> )
> >>>
> >>> // Below line works only in spark 1.3.0.
> >>> // For spark 1.1.x and spark 1.2.x,
> >>> // use bank.registerTempTable("bank") instead.
> >>> bank.toDF().registerTempTable("bank")
> >>> println(bankText.count())
> >>>
> >>> my environment are :
> >>>
> >>> spark version : 1.3.1 with hadoop 2.6
> >>>
> >>> zeppelin : binary from apache 0.5
> >>>
> >>> hadoop version : 2.6 binary from apache
> >>>
> >>> java : 1.8
> >>>
> >>> please help I am stuck here.
> >>>
> >>> thanks
> >>>
> >>> regards
> >>>
> >>> Ranveer
> >>>
> >>
> >>
> >>
                                          

Reply via email to