Hi Jonathan, Sorry or the late response, First of all thanks for your email. I'm afraid I still havent managed to get this to work. I am not clear on what you mean by register the antlr jar script though. How do I do this?
Thanks, Miguel On 23 May 2011 20:35, Jonathan Coveney <jcove...@gmail.com> wrote: > Here is what I had to do to get pig running with a different version of > Hadoop (in my case, the cloudera build but I'd try this as well): > > build pig-withouthadoop.jar by running "ant jar-withouthadoop". Then, when > you run pig, put the pig-withouthadoop.jar on your classpath as well as your > hadoop jar. In my case, I found that scripts only worked if I additionally > manually registered the antlr jar: > > register /path/to/pig/build/ivy/lib/Pig/antlr-runtime-3.2.jar; > > 2011/5/23 Rui Miguel Forte <ruimiguelfo...@gmail.com> > >> Hello, >> >> I have been trying to set up pig 0.8.1 to work with hadoop 0.20.203 >> without success. At the moment, if I run pig -x local I correctly get >> access to the grunt shell but when I try to run the default mapreduce >> option it fails with error 2999, see below: >> >> miguel@euler:~$ pig >> 2011-05-23 19:35:09,843 [main] INFO org.apache.pig.Main - Logging >> error messages to: /home/miguel/pig_1306168509838.log >> 2011-05-23 19:35:10,030 [main] INFO >> org.apache.pig.backend.hadoop.executionengine.HExecutionEngine - >> Connecting to hadoop file system at: hdfs://localhost:54310 >> 2011-05-23 19:35:10,168 [main] ERROR org.apache.pig.Main - ERROR 2999: >> Unexpected internal error. Failed to create DataStorage >> Details at logfile: /home/miguel/pig_1306168509838.log >> >> Hadoop itself runs normally on my system. Any ideas on what to do to >> fix this? Is it an incompatibility issue or something I am doing >> wrong? Below I am including my system's information, hopefully this >> will be useful for a diagnosis. >> >> The log file has the following: >> >> Error before Pig is launched >> ---------------------------- >> ERROR 2999: Unexpected internal error. Failed to create DataStorage >> >> java.lang.RuntimeException: Failed to create DataStorage >> at >> org.apache.pig.backend.hadoop.datastorage.HDataStorage.init(HDataStorage.java:75) >> at >> org.apache.pig.backend.hadoop.datastorage.HDataStorage.<init>(HDataStorage.java:58) >> at >> org.apache.pig.backend.hadoop.executionengine.HExecutionEngine.init(HExecutionEngine.java:214) >> at >> org.apache.pig.backend.hadoop.executionengine.HExecutionEngine.init(HExecutionEngine.java:134) >> at org.apache.pig.impl.PigContext.connect(PigContext.java:183) >> at org.apache.pig.PigServer.<init>(PigServer.java:226) >> at org.apache.pig.PigServer.<init>(PigServer.java:215) >> at org.apache.pig.tools.grunt.Grunt.<init>(Grunt.java:55) >> at org.apache.pig.Main.run(Main.java:452) >> at org.apache.pig.Main.main(Main.java:107) >> Caused by: java.io.IOException: Call to localhost/127.0.0.1:54310 >> failed on local exception: java.io.EOFException >> at org.apache.hadoop.ipc.Client.wrapException(Client.java:775) >> at org.apache.hadoop.ipc.Client.call(Client.java:743) >> at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:220) >> at $Proxy0.getProtocolVersion(Unknown Source) >> at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:359) >> at >> org.apache.hadoop.hdfs.DFSClient.createRPCNamenode(DFSClient.java:106) >> at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:207) >> at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:170) >> at >> org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:82) >> at >> org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1378) >> at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:66) >> at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1390) >> at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:196) >> at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:95) >> at >> org.apache.pig.backend.hadoop.datastorage.HDataStorage.init(HDataStorage.java:72) >> at >> org.apache.pig.backend.hadoop.datastorage.HDataStorage.init(HDataStorage.java:72) >> ... 9 more >> Caused by: java.io.EOFException >> at java.io.DataInputStream.readInt(DataInputStream.java:375) >> at >> org.apache.hadoop.ipc.Client$Connection.receiveResponse(Client.java:501) >> at org.apache.hadoop.ipc.Client$Connection.run(Client.java:446) >> >> ================================================================================ >> >> My system details: >> >> Ubuntu 10.04 LTS >> miguel@euler:~$ pig -version >> Apache Pig version 0.8.1 (r1094835) >> compiled Apr 18 2011, 19:26:53 >> miguel@euler:~$ hadoop version >> Hadoop 0.20.203.0 >> Subversion >> http://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20-security-203 >> -r 1099333 >> Compiled by oom on Wed May 4 07:57:50 PDT 2011 >> >> My environment variables (I know I have more than I need but tutorials >> on the web have been a little hazy on this): >> >> miguel@euler:~$ printenv | grep "PIG\|HADOOP" >> PIG_DIR=/home/miguel/pig/pig-0.8.1 >> PIG_HOME=/home/miguel/pig/pig-0.8.1 >> HADOOP_HOME=/usr/local/hadoop >> PIGDIR=/home/miguel/pig/pig-0.8.1 >> PIG_INSTALL=/home/miguel/pig/pig-0.8.1 >> PIG_HADOOP_VERSION=20 >> HADOOP_INSTALL=/usr/local/hadoop >> PIG_CLASSPATH=/usr/local/hadoop/conf >> HADOOPDIR=/usr/local/hadoop/conf >> >> Many thanks for any help, >> Rui Miguel Forte >> >