Can you check `export` to see what do you have SPARK_HOME set to?
> Date: Sun, 11 Oct 2015 16:50:23 +0530 > Subject: Re: Problem loading data from HDFS > From: ranveer.k.ku...@gmail.com > To: us...@zeppelin.incubator.apache.org > CC: dev@zeppelin.incubator.apache.org > > Still same problem my environment is now exact : > > hadoop : 2.3 > zeppelin : zeppelin-0.5.0-incubating-bin-spark-1.3.1_hadoop-2.3 > spark : spark-1.3.1-bin-hadoop2.3 > > while checking version in zeppelin using sc.version, it sowing 1.4.0 > > pls help I am stuck.. > > thanks > > On Sun, Oct 11, 2015 at 10:58 AM, Ranveer kumar <ranveer.k.ku...@gmail.com> > wrote: > > > I am using 0.5.0 now but earlier I was using own build with hadoop 2.6 > > with yarn. > > > > Is 0.5.0 binary not compatible with hadoop 2.6? > > > > I am again going to build zeppelin with hadoop 2.6 also any pre build > > zeppelin binary for hadoop 2.6 so directly download. > > > > Regards > > On 11 Oct 2015 10:07, "Felix Cheung" <felixcheun...@hotmail.com> wrote: > > > >> Are you getting the last Zeppelin release binary from > >> https://zeppelin.incubator.apache.org/download.html > >> <https://zeppelin.incubator.apache.org/download.html?> or the Apache > >> release page? > >> > >> It looks like the 0.5.0 release of Zeppelin is only built with Hadoop > >> 2.3. We might be cutting another release soon or you might need to build > >> you own... > >> > >> _____________________________ > >> From: Ranveer Kumar <ranveer.s...@gmail.com> > >> Sent: Saturday, October 10, 2015 9:00 PM > >> Subject: Re: Problem loading data from HDFS > >> To: <dev@zeppelin.incubator.apache.org> > >> Cc: <us...@zeppelin.incubator.apache.org> > >> > >> > >> Hi Felix thanks for reply. > >> > >> I am using binary download from apache. > >> > >> I also tried using 2.6 hadoop, which build from source code. > >> > >> Which version of hadoop is compatible with zeppelin binary available on > >> apache site. > >> On 11 Oct 2015 09:05, "Felix Cheung" < felixcheun...@hotmail.com> wrote: > >> > >>> Is your Zeppelin built with Hadoop 2.6? > >>> > >>> > >>> > >>> > >>> > >>> > >>> On Sat, Oct 10, 2015 at 7:35 PM -0700, "Ranveer kumar" < > >>> ranveer.k.ku...@gmail.com> wrote: > >>> Hi All, > >>> > >>> I am new in Zepplin and HDFS. I manage to install zeppelin and working > >>> fine > >>> while loading data from local directory . But when same I am trying to > >>> load > >>> from HDFS (install locally standalone mode). > >>> > >>> here is my code : > >>> > >>> val bankText = sc.textFile("/home/ranveer/Desktop/CSVs/bank-full.csv") > >>> > >>> above is working fine. > >>> > >>> but when trying form hdfs : > >>> > >>> val bankText = sc.textFile("hdfs:// 127.0.0.1:9000/demo/ > >>> < http://127.0.0.1:9000/demo/csv/bank-full.csv>csv > >>> < http://127.0.0.1:9000/demo/csv/bank-full.csv>/ > >>> < http://127.0.0.1:9000/demo/csv/bank-full.csv>bank-full.csv > >>> < http://127.0.0.1:9000/demo/csv/bank-full.csv>") > >>> > >>> not working and giving error : > >>> > >>> java.lang.VerifyError: class > >>> org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$GetAdditionalDatanodeRequestProto > >>> > >>> overrides final method > >>> getUnknownFields.()Lcom/google/protobuf/UnknownFieldSet; at > >>> java.lang.ClassLoader.defineClass1(Native Method) at > >>> java.lang.ClassLoader.defineClass(ClassLoader.java:800) at > >>> java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142) > >>> at > >>> java.net.URLClassLoader.defineClass(URLClassLoader.java:449) at > >>> java.net.URLClassLoader.access$100(URLClassLoader.java:71) > >>> > >>> complete code is : > >>> > >>> val bankText = sc.textFile("hdfs:// 127.0.0.1:9000/demo/ > >>> < http://127.0.0.1:9000/demo/csv/bank-full.csv>csv > >>> < http://127.0.0.1:9000/demo/csv/bank-full.csv>/ > >>> < http://127.0.0.1:9000/demo/csv/bank-full.csv>bank-full.csv > >>> < http://127.0.0.1:9000/demo/csv/bank-full.csv>") > >>> // val bankText = > >>> sc.textFile("/home/ranveer/Desktop/CSVs/bank-full.csv") > >>> case class Bank(age:Integer, job:String, marital : String, education : > >>> String, balance : Integer) > >>> > >>> val bank = bankText.map(s=>s.split(";")).filter(s=>s(0)!="\"age\"").map( > >>> s=>Bank(s(0).toInt, > >>> s(1).replaceAll("\"", ""), > >>> s(2).replaceAll("\"", ""), > >>> s(3).replaceAll("\"", ""), > >>> s(5).replaceAll("\"", "").toInt > >>> ) > >>> ) > >>> > >>> // Below line works only in spark 1.3.0. > >>> // For spark 1.1.x and spark 1.2.x, > >>> // use bank.registerTempTable("bank") instead. > >>> bank.toDF().registerTempTable("bank") > >>> println(bankText.count()) > >>> > >>> my environment are : > >>> > >>> spark version : 1.3.1 with hadoop 2.6 > >>> > >>> zeppelin : binary from apache 0.5 > >>> > >>> hadoop version : 2.6 binary from apache > >>> > >>> java : 1.8 > >>> > >>> please help I am stuck here. > >>> > >>> thanks > >>> > >>> regards > >>> > >>> Ranveer > >>> > >> > >> > >>