I am pasting from a response ot another user...I think it may apply

"Here is what I had to do to get pig running with a different version of
Hadoop (in my case, the cloudera build but I'd try this as well):

build pig-withouthadoop.jar by running "ant jar-withouthadoop". Then, when
you run pig, put the pig-withouthadoop.jar on your classpath as well as your
hadoop jar. In my case, I found that scripts only worked if I additionally
manually registered the antlr jar:

register /path/to/pig/build/ivy/lib/Pig/antlr-runtime-3.2.jar;"



2011/5/25 Mohit Anchlia <[email protected]>

> On Wed, May 25, 2011 at 5:02 PM, Mohit Anchlia <[email protected]>
> wrote:
> > I am in process of installing and learning pig. I have a hadoop
> > cluster and when I try to run pig in mapreduce mode it errors out:
>
> Hadoop version is hadoop-0.20.203.0 and pig version is pig-0.8.1
>
> >
> > Error before Pig is launched
> > ----------------------------
> > ERROR 2999: Unexpected internal error. Failed to create DataStorage
> >
> > java.lang.RuntimeException: Failed to create DataStorage
> >        at
> org.apache.pig.backend.hadoop.datastorage.HDataStorage.init(HDataStorage.java:75)
> >        at
> org.apache.pig.backend.hadoop.datastorage.HDataStorage.<init>(HDataStorage.java:58)
> >        at
> org.apache.pig.backend.hadoop.executionengine.HExecutionEngine.init(HExecutionEngine.java:214)
> >        at
> org.apache.pig.backend.hadoop.executionengine.HExecutionEngine.init(HExecutionEngine.java:134)
> >        at org.apache.pig.impl.PigContext.connect(PigContext.java:183)
> >        at org.apache.pig.PigServer.<init>(PigServer.java:226)
> >        at org.apache.pig.PigServer.<init>(PigServer.java:215)
> >        at org.apache.pig.tools.grunt.Grunt.<init>(Grunt.java:55)
> >        at org.apache.pig.Main.run(Main.java:452)
> >        at org.apache.pig.Main.main(Main.java:107)
> > Caused by: java.io.IOException: Call to dsdb1/172.18.60.96:54310
> > failed on local exception: java.io.EOFException
> >        at org.apache.hadoop.ipc.Client.wrapException(Client.java:775)
> >        at org.apache.hadoop.ipc.Client.call(Client.java:743)
> >        at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:220)
> >        at $Proxy0.getProtocolVersion(Unknown Source)
> >        at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:359)
> >        at
> org.apache.hadoop.hdfs.DFSClient.createRPCNamenode(DFSClient.java:106)
> >        at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:207)
> >        at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:170)
> >        at
> org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:82)
> >        at
> org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1378)
> >        at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:66)
> >        at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1390)
> >        at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:196)
> >        at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:95)
> >        at
> org.apache.pig.backend.hadoop.datastorage.HDataStorage.init(HDataStorage.java:72)
> >        ... 9 more
> > Caused by: java.io.EOFException
> >        at java.io.DataInputStream.readInt(DataInputStream.java:375)
> >        at
> org.apache.hadoop.ipc.Client$Connection.receiveResponse(Client.java:501)
> >        at org.apache.hadoop.ipc.Client$Connection.run(Client.java:446)
> >
>

Reply via email to