Are you getting the last Zeppelin release binary from https://zeppelin.incubator.apache.org/download.html or the Apache release page? It looks like the 0.5.0 release of Zeppelin is only built with Hadoop 2.3. We might be cutting another release soon or you might need to build you own...
_____________________________ From: Ranveer Kumar <ranveer.s...@gmail.com> Sent: Saturday, October 10, 2015 9:00 PM Subject: Re: Problem loading data from HDFS To: <d...@zeppelin.incubator.apache.org> Cc: <users@zeppelin.incubator.apache.org> Hi Felix thanks for reply. I am using binary download from apache. I also tried using 2.6 hadoop, which build from source code. Which version of hadoop is compatible with zeppelin binary available on apache site. On 11 Oct 2015 09:05, "Felix Cheung" < felixcheun...@hotmail.com> wrote: Is your Zeppelin built with Hadoop 2.6? On Sat, Oct 10, 2015 at 7:35 PM -0700, "Ranveer kumar" < ranveer.k.ku...@gmail.com> wrote: Hi All, I am new in Zepplin and HDFS. I manage to install zeppelin and working fine while loading data from local directory . But when same I am trying to load from HDFS (install locally standalone mode). here is my code : val bankText = sc.textFile("/home/ranveer/Desktop/CSVs/bank-full.csv") above is working fine. but when trying form hdfs : val bankText = sc.textFile("hdfs:// 127.0.0.1:9000/demo/ < http://127.0.0.1:9000/demo/csv/bank-full.csv>csv < http://127.0.0.1:9000/demo/csv/bank-full.csv>/ < http://127.0.0.1:9000/demo/csv/bank-full.csv>bank-full.csv < http://127.0.0.1:9000/demo/csv/bank-full.csv>") not working and giving error : java.lang.VerifyError: class org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$GetAdditionalDatanodeRequestProto overrides final method getUnknownFields.()Lcom/google/protobuf/UnknownFieldSet; at java.lang.ClassLoader.defineClass1(Native Method) at java.lang.ClassLoader.defineClass(ClassLoader.java:800) at java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142) at java.net.URLClassLoader.defineClass(URLClassLoader.java:449) at java.net.URLClassLoader.access$100(URLClassLoader.java:71) complete code is : val bankText = sc.textFile("hdfs:// 127.0.0.1:9000/demo/ < http://127.0.0.1:9000/demo/csv/bank-full.csv>csv < http://127.0.0.1:9000/demo/csv/bank-full.csv>/ < http://127.0.0.1:9000/demo/csv/bank-full.csv>bank-full.csv < http://127.0.0.1:9000/demo/csv/bank-full.csv>") // val bankText = sc.textFile("/home/ranveer/Desktop/CSVs/bank-full.csv") case class Bank(age:Integer, job:String, marital : String, education : String, balance : Integer) val bank = bankText.map(s=>s.split(";")).filter(s=>s(0)!="\"age\"").map( s=>Bank(s(0).toInt, s(1).replaceAll("\"", ""), s(2).replaceAll("\"", ""), s(3).replaceAll("\"", ""), s(5).replaceAll("\"", "").toInt ) ) // Below line works only in spark 1.3.0. // For spark 1.1.x and spark 1.2.x, // use bank.registerTempTable("bank") instead. bank.toDF().registerTempTable("bank") println(bankText.count()) my environment are : spark version : 1.3.1 with hadoop 2.6 zeppelin : binary from apache 0.5 hadoop version : 2.6 binary from apache java : 1.8 please help I am stuck here. thanks regards Ranveer