Nice spot on, but it seems that's not the problem...I am able to run this successfully on pig 0.9 and 0.10 any ideas how further I can debug the problem?
I also double checked all the variables, seems to be fine: PIG_CLASSPATH=/opt/hcat/share/hcatalog/hcatalog-0.4.0.jar:/usr/lib/hive/conf:/usr/lib/hadoop-0.20/conf /usr/lib/hive/conf and /usr/lib/hadoop-0.20/conf also exist. Thanks. On 12 November 2012 17:37, Cheolsoo Park <[email protected]> wrote: > Hi Michal, > > Caused by: java.lang.IllegalArgumentException: Can not create a Path > from an empty string > at org.apache.hadoop.fs.Path.checkPathArg(Path.java:82) > at org.apache.hadoop.fs.Path.<init>(Path.java:90) > at org.apache.hadoop.fs.Path.<init>(Path.java:45 > > Your error message indicates that there is a typo somewhere in paths. I > believe that your PIG_CLASSPATH is the problem: > > > PIG_CLASSPATH=$HCAT_HOME/share/hcatalog/hcatalog-0.4.0.jar::$HIVE_HOME/conf:$HADOOP_HOME/conf > > You have a double colon :: in the middle, and that will be interpreted as > an empty string. > > Thanks, > Cheolsoo > > On Mon, Nov 12, 2012 at 8:47 AM, Michał Czerwiński < > [email protected] > > wrote: > > > I am trying to use pig 0.11 and pig trunk (currently 0.12) because pig > 0.10 > > seems to be having issues with python udf... > > > > According to this > > http://www.mail-archive.com/[email protected]/msg05837.html > > > > " after replacing pig.jar and pig-withouthadoop.jar with the > > 0.11 ones from the svn trunk, they work like a charm." > > > > Well this is clearly not the case for me... > > > > The error I get is: > > > > Pig Stack Trace > > --------------- > > ERROR 2017: Internal error creating job configuration. > > > > org.apache.pig.impl.logicalLayer.FrontendException: ERROR 1066: Unable to > > open iterator for alias ll > > at org.apache.pig.PigServer.openIterator(PigServer.java:841) > > at > org.apache.pig.tools.grunt.GruntParser.processDump(GruntParser.java:696) > > at > > > > > org.apache.pig.tools.pigscript.parser.PigScriptParser.parse(PigScriptParser.java:320) > > at > > > > > org.apache.pig.tools.grunt.GruntParser.parseStopOnError(GruntParser.java:194) > > at > > > > > org.apache.pig.tools.grunt.GruntParser.parseStopOnError(GruntParser.java:170) > > at org.apache.pig.tools.grunt.Grunt.run(Grunt.java:69) > > at org.apache.pig.Main.run(Main.java:535) > > at org.apache.pig.Main.main(Main.java:154) > > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > > at > > > > > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) > > at > > > > > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) > > at java.lang.reflect.Method.invoke(Method.java:597) > > at org.apache.hadoop.util.RunJar.main(RunJar.java:197) > > Caused by: org.apache.pig.PigException: ERROR 1002: Unable to store alias > > ll > > at org.apache.pig.PigServer.storeEx(PigServer.java:940) > > at org.apache.pig.PigServer.store(PigServer.java:903) > > at org.apache.pig.PigServer.openIterator(PigServer.java:816) > > ... 12 more > > Caused by: > > > > > org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.JobCreationException: > > ERROR 2017: Internal error creating job configuration. > > at > > > > > org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.JobControlCompiler.getJob(JobControlCompiler.java:848) > > at > > > > > org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.JobControlCompiler.compile(JobControlCompiler.java:294) > > at > > > > > org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher.launchPig(MapReduceLauncher.java:177) > > at org.apache.pig.PigServer.launchPlan(PigServer.java:1269) > > at > > org.apache.pig.PigServer.executeCompiledLogicalPlan(PigServer.java:1254) > > at org.apache.pig.PigServer.storeEx(PigServer.java:936) > > ... 14 more > > Caused by: java.lang.IllegalArgumentException: Can not create a Path from > > an empty string > > at org.apache.hadoop.fs.Path.checkPathArg(Path.java:82) > > at org.apache.hadoop.fs.Path.<init>(Path.java:90) > > at org.apache.hadoop.fs.Path.<init>(Path.java:45) > > at > > > > > org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.JobControlCompiler.shipToHDFS(JobControlCompiler.java:1455) > > at > > > > > org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.JobControlCompiler.putJarOnClassPathThroughDistributedCache(JobControlCompiler.java:1432) > > at > > > > > org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.JobControlCompiler.getJob(JobControlCompiler.java:508) > > ... 19 more > > > > > ================================================================================ > > > > I am using the following startup script: > > > > export HADOOP_HOME=/usr/lib/hadoop-0.20 > > export HCAT_HOME=/opt/hcat > > export HIVE_HOME=/usr/lib/hive > > > > > > > PIG_CLASSPATH=$HCAT_HOME/share/hcatalog/hcatalog-0.4.0.jar::$HIVE_HOME/conf:$HADOOP_HOME/conf > > > > for file in $HIVE_HOME/lib/*.jar; do > > echo "==> Adding $file" > > PIG_CLASSPATH=$PIG_CLASSPATH:$file > > done > > > > export PIG_OPTS=-Dhive.metastore.uris=thrift:// > > appserver.hadoop.staging.qutics.com:10002 > > export JAVA_HOME=$(readlink -f /usr/bin/java | sed "s:bin/java::") > > exec bin/pig -Dpig.additional.jars=$PIG_CLASSPATH "$@" > > > > Any clues? > > Thank you! > > >
