Jeff, I am unable to reproduce this error using pig compiled from the current top of the 0.6 branch and the script you provided. Are you sure 0.6 is what you are actually using? It hasn't been released yet. Do you know what svn revision the jar was compiled from?
-D On Sat, Jan 9, 2010 at 5:07 PM, Dmitriy Ryaboy <[email protected]> wrote: > Jeff, > I'll check it out this weekend. > > -D > > On Sat, Jan 9, 2010 at 3:47 PM, Jeff Dalton <[email protected]> wrote: >> I downloaded the version of PiggyBank from the 0.6 branch, compiled, >> and deployed it. However, I still get the same error message: >> >> java.lang.ClassCastException: org.apache.pig.ExecType cannot be cast >> to org.apache.pig.impl.PigContext >> at >> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.SliceWrapper.readFields(SliceWrapper.java:168) >> at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:333) >> at org.apache.hadoop.mapred.MapTask.run(MapTask.java:307) >> at org.apache.hadoop.mapred.Child.main(Child.java:159) >> >> I'll try again later, but if anyone has any insights, I would >> appreciate the help. >> >> Thanks, >> >> - Jeff >> >> On Sat, Jan 9, 2010 at 6:33 PM, Jeff Dalton <[email protected]> wrote: >>> Ahh, the PiggyBank version was the latest from Trunk. I probably need >>> to go track down the version from the 0.6 branch. >>> >>> On Sat, Jan 9, 2010 at 6:26 PM, Dmitriy Ryaboy <[email protected]> wrote: >>>> When you say that the code is from SVN, do you mean trunk, or the 0.6 >>>> branch? >>>> >>>> >>>> On Sat, Jan 9, 2010 at 3:22 PM, Jeff Dalton <[email protected]> >>>> wrote: >>>>> A cluster I'm using was recently upgraded to PIG 0.6. Since then, >>>>> I've been having problems with scripts that use PiggyBank functions. >>>>> All the map jobs for the script fail with: >>>>> WARN org.apache.hadoop.mapred.Child: Error running child >>>>> java.lang.ClassCastException: org.apache.pig.ExecType cannot be cast >>>>> to org.apache.pig.impl.PigContext >>>>> at >>>>> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.SliceWrapper.readFields(SliceWrapper.java:168) >>>>> at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:333) >>>>> at org.apache.hadoop.mapred.MapTask.run(MapTask.java:307) >>>>> at org.apache.hadoop.mapred.Child.main(Child.java:159) >>>>> INFO org.apache.hadoop.mapred.Task: Runnning cleanup for the task >>>>> >>>>> I compiled the PiggyBank jar using the latest code from SVN (as of Jan >>>>> 9) and Pig 0.6. Below I've included a simple example program that >>>>> caused the error that simply reads a text file of words and lower >>>>> cases them. >>>>> >>>>> register ./piggybank.jar >>>>> DEFINE ToLower org.apache.pig.piggybank.evaluation.string.LOWER(); >>>>> words = LOAD './data/headwords_sample' USING PigStorage() as >>>>> (word:charArray); >>>>> lowerCaseWords = FOREACH words GENERATE ToLower(word) as word; >>>>> STORE lowerCaseWords into './tmp/cooc3' USING PigStorage(); >>>>> >>>>> The Hadoop error isn't very informative about what is going on. Am I >>>>> using a compatible version of PiggyBank? What should I be doing >>>>> differently? >>>>> >>>>> Thanks, >>>>> >>>>> - Jeff >>>>> >>>> >>> >> >
