I am using 0.5.0 now but earlier I was using own build with hadoop 2.6 with
yarn.

Is 0.5.0 binary not compatible with hadoop 2.6?

I am again going to build zeppelin with hadoop 2.6 also any pre build
zeppelin binary for hadoop 2.6 so directly download.

Regards
On 11 Oct 2015 10:07, "Felix Cheung" <felixcheun...@hotmail.com> wrote:

> Are you getting the last Zeppelin release binary from
> https://zeppelin.incubator.apache.org/download.html
> <https://zeppelin.incubator.apache.org/download.html?> or the Apache
> release page?
>
> It looks like the 0.5.0 release of Zeppelin is only built with Hadoop 2.3.
> We might be cutting another release soon or you might need to build you
> own...
>
> _____________________________
> From: Ranveer Kumar <ranveer.s...@gmail.com>
> Sent: Saturday, October 10, 2015 9:00 PM
> Subject: Re: Problem loading data from HDFS
> To: <dev@zeppelin.incubator.apache.org>
> Cc: <us...@zeppelin.incubator.apache.org>
>
>
> Hi Felix thanks for reply.
>
> I am using binary download from apache.
>
> I also tried using 2.6 hadoop, which build from source code.
>
> Which version of hadoop is compatible with zeppelin binary available on
> apache site.
> On 11 Oct 2015 09:05, "Felix Cheung" < felixcheun...@hotmail.com> wrote:
>
>> Is your Zeppelin built with Hadoop 2.6?
>>
>>
>>
>>
>>
>>
>> On Sat, Oct 10, 2015 at 7:35 PM -0700, "Ranveer kumar" <
>> ranveer.k.ku...@gmail.com> wrote:
>> Hi All,
>>
>> I am new in Zepplin and HDFS. I manage to install zeppelin and working
>> fine
>> while loading data from local directory . But when same I am trying to
>> load
>> from HDFS (install locally standalone mode).
>>
>> here is my code :
>>
>> val bankText = sc.textFile("/home/ranveer/Desktop/CSVs/bank-full.csv")
>>
>> above is working fine.
>>
>> but when trying form hdfs :
>>
>> val bankText = sc.textFile("hdfs:// 127.0.0.1:9000/demo/
>> < http://127.0.0.1:9000/demo/csv/bank-full.csv>csv
>> < http://127.0.0.1:9000/demo/csv/bank-full.csv>/
>> < http://127.0.0.1:9000/demo/csv/bank-full.csv>bank-full.csv
>> < http://127.0.0.1:9000/demo/csv/bank-full.csv>")
>>
>> not working and giving error :
>>
>> java.lang.VerifyError: class
>> org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$GetAdditionalDatanodeRequestProto
>>
>> overrides final method
>> getUnknownFields.()Lcom/google/protobuf/UnknownFieldSet; at
>> java.lang.ClassLoader.defineClass1(Native Method) at
>> java.lang.ClassLoader.defineClass(ClassLoader.java:800) at
>> java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142)
>> at
>> java.net.URLClassLoader.defineClass(URLClassLoader.java:449) at
>> java.net.URLClassLoader.access$100(URLClassLoader.java:71)
>>
>> complete code is :
>>
>> val bankText = sc.textFile("hdfs:// 127.0.0.1:9000/demo/
>> < http://127.0.0.1:9000/demo/csv/bank-full.csv>csv
>> < http://127.0.0.1:9000/demo/csv/bank-full.csv>/
>> < http://127.0.0.1:9000/demo/csv/bank-full.csv>bank-full.csv
>> < http://127.0.0.1:9000/demo/csv/bank-full.csv>")
>> // val bankText = sc.textFile("/home/ranveer/Desktop/CSVs/bank-full.csv")
>> case class Bank(age:Integer, job:String, marital : String, education :
>> String, balance : Integer)
>>
>> val bank = bankText.map(s=>s.split(";")).filter(s=>s(0)!="\"age\"").map(
>>     s=>Bank(s(0).toInt,
>>             s(1).replaceAll("\"", ""),
>>             s(2).replaceAll("\"", ""),
>>             s(3).replaceAll("\"", ""),
>>             s(5).replaceAll("\"", "").toInt
>>         )
>> )
>>
>> // Below line works only in spark 1.3.0.
>> // For spark 1.1.x and spark 1.2.x,
>> // use bank.registerTempTable("bank") instead.
>> bank.toDF().registerTempTable("bank")
>> println(bankText.count())
>>
>> my environment are :
>>
>> spark version : 1.3.1 with hadoop 2.6
>>
>> zeppelin : binary from apache 0.5
>>
>> hadoop version : 2.6 binary from apache
>>
>> java : 1.8
>>
>> please help I am stuck here.
>>
>> thanks
>>
>> regards
>>
>> Ranveer
>>
>
>
>

Reply via email to