I am pasting from a response ot another user...I think it may apply "Here is what I had to do to get pig running with a different version of Hadoop (in my case, the cloudera build but I'd try this as well):
build pig-withouthadoop.jar by running "ant jar-withouthadoop". Then, when you run pig, put the pig-withouthadoop.jar on your classpath as well as your hadoop jar. In my case, I found that scripts only worked if I additionally manually registered the antlr jar: register /path/to/pig/build/ivy/lib/Pig/antlr-runtime-3.2.jar;" 2011/5/25 Mohit Anchlia <[email protected]> > On Wed, May 25, 2011 at 5:02 PM, Mohit Anchlia <[email protected]> > wrote: > > I am in process of installing and learning pig. I have a hadoop > > cluster and when I try to run pig in mapreduce mode it errors out: > > Hadoop version is hadoop-0.20.203.0 and pig version is pig-0.8.1 > > > > > Error before Pig is launched > > ---------------------------- > > ERROR 2999: Unexpected internal error. Failed to create DataStorage > > > > java.lang.RuntimeException: Failed to create DataStorage > > at > org.apache.pig.backend.hadoop.datastorage.HDataStorage.init(HDataStorage.java:75) > > at > org.apache.pig.backend.hadoop.datastorage.HDataStorage.<init>(HDataStorage.java:58) > > at > org.apache.pig.backend.hadoop.executionengine.HExecutionEngine.init(HExecutionEngine.java:214) > > at > org.apache.pig.backend.hadoop.executionengine.HExecutionEngine.init(HExecutionEngine.java:134) > > at org.apache.pig.impl.PigContext.connect(PigContext.java:183) > > at org.apache.pig.PigServer.<init>(PigServer.java:226) > > at org.apache.pig.PigServer.<init>(PigServer.java:215) > > at org.apache.pig.tools.grunt.Grunt.<init>(Grunt.java:55) > > at org.apache.pig.Main.run(Main.java:452) > > at org.apache.pig.Main.main(Main.java:107) > > Caused by: java.io.IOException: Call to dsdb1/172.18.60.96:54310 > > failed on local exception: java.io.EOFException > > at org.apache.hadoop.ipc.Client.wrapException(Client.java:775) > > at org.apache.hadoop.ipc.Client.call(Client.java:743) > > at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:220) > > at $Proxy0.getProtocolVersion(Unknown Source) > > at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:359) > > at > org.apache.hadoop.hdfs.DFSClient.createRPCNamenode(DFSClient.java:106) > > at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:207) > > at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:170) > > at > org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:82) > > at > org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1378) > > at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:66) > > at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1390) > > at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:196) > > at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:95) > > at > org.apache.pig.backend.hadoop.datastorage.HDataStorage.init(HDataStorage.java:72) > > ... 9 more > > Caused by: java.io.EOFException > > at java.io.DataInputStream.readInt(DataInputStream.java:375) > > at > org.apache.hadoop.ipc.Client$Connection.receiveResponse(Client.java:501) > > at org.apache.hadoop.ipc.Client$Connection.run(Client.java:446) > > >
