Jeff, I'll check it out this weekend. -D
On Sat, Jan 9, 2010 at 3:47 PM, Jeff Dalton <[email protected]> wrote: > I downloaded the version of PiggyBank from the 0.6 branch, compiled, > and deployed it. However, I still get the same error message: > > java.lang.ClassCastException: org.apache.pig.ExecType cannot be cast > to org.apache.pig.impl.PigContext > at > org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.SliceWrapper.readFields(SliceWrapper.java:168) > at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:333) > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:307) > at org.apache.hadoop.mapred.Child.main(Child.java:159) > > I'll try again later, but if anyone has any insights, I would > appreciate the help. > > Thanks, > > - Jeff > > On Sat, Jan 9, 2010 at 6:33 PM, Jeff Dalton <[email protected]> wrote: >> Ahh, the PiggyBank version was the latest from Trunk. I probably need >> to go track down the version from the 0.6 branch. >> >> On Sat, Jan 9, 2010 at 6:26 PM, Dmitriy Ryaboy <[email protected]> wrote: >>> When you say that the code is from SVN, do you mean trunk, or the 0.6 >>> branch? >>> >>> >>> On Sat, Jan 9, 2010 at 3:22 PM, Jeff Dalton <[email protected]> wrote: >>>> A cluster I'm using was recently upgraded to PIG 0.6. Since then, >>>> I've been having problems with scripts that use PiggyBank functions. >>>> All the map jobs for the script fail with: >>>> WARN org.apache.hadoop.mapred.Child: Error running child >>>> java.lang.ClassCastException: org.apache.pig.ExecType cannot be cast >>>> to org.apache.pig.impl.PigContext >>>> at >>>> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.SliceWrapper.readFields(SliceWrapper.java:168) >>>> at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:333) >>>> at org.apache.hadoop.mapred.MapTask.run(MapTask.java:307) >>>> at org.apache.hadoop.mapred.Child.main(Child.java:159) >>>> INFO org.apache.hadoop.mapred.Task: Runnning cleanup for the task >>>> >>>> I compiled the PiggyBank jar using the latest code from SVN (as of Jan >>>> 9) and Pig 0.6. Below I've included a simple example program that >>>> caused the error that simply reads a text file of words and lower >>>> cases them. >>>> >>>> register ./piggybank.jar >>>> DEFINE ToLower org.apache.pig.piggybank.evaluation.string.LOWER(); >>>> words = LOAD './data/headwords_sample' USING PigStorage() as >>>> (word:charArray); >>>> lowerCaseWords = FOREACH words GENERATE ToLower(word) as word; >>>> STORE lowerCaseWords into './tmp/cooc3' USING PigStorage(); >>>> >>>> The Hadoop error isn't very informative about what is going on. Am I >>>> using a compatible version of PiggyBank? What should I be doing >>>> differently? >>>> >>>> Thanks, >>>> >>>> - Jeff >>>> >>> >> >
